var/home/core/zuul-output/0000755000175000017500000000000015117775520014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015120011524015460 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005062126415120011515017672 0ustar rootrootDec 15 12:17:17 crc systemd[1]: Starting Kubernetes Kubelet... Dec 15 12:17:17 crc restorecon[4581]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:17 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:18 crc restorecon[4581]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 15 12:17:18 crc restorecon[4581]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 15 12:17:19 crc kubenswrapper[4719]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 15 12:17:19 crc kubenswrapper[4719]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 15 12:17:19 crc kubenswrapper[4719]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 15 12:17:19 crc kubenswrapper[4719]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 15 12:17:19 crc kubenswrapper[4719]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 15 12:17:19 crc kubenswrapper[4719]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.322774 4719 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326021 4719 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326042 4719 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326047 4719 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326051 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326057 4719 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326063 4719 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326069 4719 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326077 4719 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326084 4719 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326091 4719 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326098 4719 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326104 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326109 4719 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326116 4719 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326122 4719 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326127 4719 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326139 4719 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326143 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326147 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326151 4719 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326155 4719 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326159 4719 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326163 4719 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326167 4719 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326170 4719 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326174 4719 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326178 4719 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326181 4719 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326185 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326188 4719 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326192 4719 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326196 4719 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326230 4719 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326235 4719 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326238 4719 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326242 4719 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326246 4719 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326250 4719 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326255 4719 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326258 4719 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326263 4719 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326267 4719 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326270 4719 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326274 4719 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326279 4719 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326282 4719 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326287 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326291 4719 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326295 4719 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326300 4719 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326306 4719 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326312 4719 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326316 4719 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326320 4719 feature_gate.go:330] unrecognized feature gate: Example Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326324 4719 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326328 4719 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326331 4719 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326335 4719 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326339 4719 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326342 4719 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326345 4719 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326349 4719 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326352 4719 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326356 4719 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326360 4719 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326365 4719 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326370 4719 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326374 4719 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326378 4719 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326381 4719 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.326384 4719 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326789 4719 flags.go:64] FLAG: --address="0.0.0.0" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326805 4719 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326818 4719 flags.go:64] FLAG: --anonymous-auth="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326827 4719 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326836 4719 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326842 4719 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326851 4719 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326881 4719 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326888 4719 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326892 4719 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326898 4719 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326903 4719 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326908 4719 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326912 4719 flags.go:64] FLAG: --cgroup-root="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326917 4719 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326921 4719 flags.go:64] FLAG: --client-ca-file="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326927 4719 flags.go:64] FLAG: --cloud-config="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326931 4719 flags.go:64] FLAG: --cloud-provider="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326935 4719 flags.go:64] FLAG: --cluster-dns="[]" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326942 4719 flags.go:64] FLAG: --cluster-domain="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326946 4719 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326951 4719 flags.go:64] FLAG: --config-dir="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326955 4719 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326961 4719 flags.go:64] FLAG: --container-log-max-files="5" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326966 4719 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326971 4719 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326975 4719 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326980 4719 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326984 4719 flags.go:64] FLAG: --contention-profiling="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326988 4719 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326992 4719 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.326997 4719 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327002 4719 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327007 4719 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327012 4719 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327016 4719 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327020 4719 flags.go:64] FLAG: --enable-load-reader="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327024 4719 flags.go:64] FLAG: --enable-server="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327028 4719 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327035 4719 flags.go:64] FLAG: --event-burst="100" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327040 4719 flags.go:64] FLAG: --event-qps="50" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327044 4719 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327049 4719 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327053 4719 flags.go:64] FLAG: --eviction-hard="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327060 4719 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327064 4719 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327069 4719 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327073 4719 flags.go:64] FLAG: --eviction-soft="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327077 4719 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327082 4719 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327086 4719 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327091 4719 flags.go:64] FLAG: --experimental-mounter-path="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327095 4719 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327100 4719 flags.go:64] FLAG: --fail-swap-on="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327105 4719 flags.go:64] FLAG: --feature-gates="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327112 4719 flags.go:64] FLAG: --file-check-frequency="20s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327118 4719 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327123 4719 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327129 4719 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327135 4719 flags.go:64] FLAG: --healthz-port="10248" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327140 4719 flags.go:64] FLAG: --help="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327146 4719 flags.go:64] FLAG: --hostname-override="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327151 4719 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327157 4719 flags.go:64] FLAG: --http-check-frequency="20s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327163 4719 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327169 4719 flags.go:64] FLAG: --image-credential-provider-config="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327174 4719 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327180 4719 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327185 4719 flags.go:64] FLAG: --image-service-endpoint="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327190 4719 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327195 4719 flags.go:64] FLAG: --kube-api-burst="100" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327199 4719 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327204 4719 flags.go:64] FLAG: --kube-api-qps="50" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327208 4719 flags.go:64] FLAG: --kube-reserved="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327213 4719 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327217 4719 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327221 4719 flags.go:64] FLAG: --kubelet-cgroups="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327225 4719 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327230 4719 flags.go:64] FLAG: --lock-file="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327234 4719 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327238 4719 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327242 4719 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327249 4719 flags.go:64] FLAG: --log-json-split-stream="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327253 4719 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327257 4719 flags.go:64] FLAG: --log-text-split-stream="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327262 4719 flags.go:64] FLAG: --logging-format="text" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327266 4719 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327271 4719 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327275 4719 flags.go:64] FLAG: --manifest-url="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327280 4719 flags.go:64] FLAG: --manifest-url-header="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327285 4719 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327290 4719 flags.go:64] FLAG: --max-open-files="1000000" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327296 4719 flags.go:64] FLAG: --max-pods="110" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327301 4719 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327305 4719 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327310 4719 flags.go:64] FLAG: --memory-manager-policy="None" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327315 4719 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327319 4719 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327323 4719 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327327 4719 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327340 4719 flags.go:64] FLAG: --node-status-max-images="50" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327344 4719 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327348 4719 flags.go:64] FLAG: --oom-score-adj="-999" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327361 4719 flags.go:64] FLAG: --pod-cidr="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327365 4719 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327372 4719 flags.go:64] FLAG: --pod-manifest-path="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327376 4719 flags.go:64] FLAG: --pod-max-pids="-1" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327381 4719 flags.go:64] FLAG: --pods-per-core="0" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327385 4719 flags.go:64] FLAG: --port="10250" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327389 4719 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327393 4719 flags.go:64] FLAG: --provider-id="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327397 4719 flags.go:64] FLAG: --qos-reserved="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327402 4719 flags.go:64] FLAG: --read-only-port="10255" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327406 4719 flags.go:64] FLAG: --register-node="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327410 4719 flags.go:64] FLAG: --register-schedulable="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327414 4719 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327422 4719 flags.go:64] FLAG: --registry-burst="10" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327426 4719 flags.go:64] FLAG: --registry-qps="5" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327431 4719 flags.go:64] FLAG: --reserved-cpus="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327435 4719 flags.go:64] FLAG: --reserved-memory="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327441 4719 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327445 4719 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327449 4719 flags.go:64] FLAG: --rotate-certificates="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327453 4719 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327457 4719 flags.go:64] FLAG: --runonce="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327461 4719 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327466 4719 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327471 4719 flags.go:64] FLAG: --seccomp-default="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327475 4719 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327479 4719 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327483 4719 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327488 4719 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327493 4719 flags.go:64] FLAG: --storage-driver-password="root" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327497 4719 flags.go:64] FLAG: --storage-driver-secure="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327501 4719 flags.go:64] FLAG: --storage-driver-table="stats" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327505 4719 flags.go:64] FLAG: --storage-driver-user="root" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327509 4719 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327513 4719 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327518 4719 flags.go:64] FLAG: --system-cgroups="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327523 4719 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327530 4719 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327534 4719 flags.go:64] FLAG: --tls-cert-file="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327538 4719 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327546 4719 flags.go:64] FLAG: --tls-min-version="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327550 4719 flags.go:64] FLAG: --tls-private-key-file="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327555 4719 flags.go:64] FLAG: --topology-manager-policy="none" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327559 4719 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327563 4719 flags.go:64] FLAG: --topology-manager-scope="container" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327567 4719 flags.go:64] FLAG: --v="2" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327573 4719 flags.go:64] FLAG: --version="false" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327580 4719 flags.go:64] FLAG: --vmodule="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327585 4719 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.327590 4719 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327690 4719 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327695 4719 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327700 4719 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327703 4719 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327707 4719 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327711 4719 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327716 4719 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327720 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327724 4719 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327729 4719 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327733 4719 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327737 4719 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327740 4719 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327745 4719 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327749 4719 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327753 4719 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327757 4719 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327761 4719 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327765 4719 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327769 4719 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327772 4719 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327777 4719 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327787 4719 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327791 4719 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327797 4719 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327802 4719 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327807 4719 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327812 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327817 4719 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327822 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327826 4719 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327829 4719 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327835 4719 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327839 4719 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327842 4719 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327872 4719 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327877 4719 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327882 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327886 4719 feature_gate.go:330] unrecognized feature gate: Example Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327891 4719 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327895 4719 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327900 4719 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327904 4719 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327908 4719 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327913 4719 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327917 4719 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327921 4719 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327924 4719 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327928 4719 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327931 4719 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327935 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327939 4719 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327942 4719 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327946 4719 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327950 4719 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327953 4719 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327957 4719 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327960 4719 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327972 4719 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327976 4719 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327979 4719 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327983 4719 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327986 4719 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327990 4719 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327994 4719 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.327997 4719 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.328001 4719 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.328006 4719 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.328009 4719 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.328013 4719 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.328017 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.328023 4719 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.337355 4719 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.337398 4719 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337496 4719 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337506 4719 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337510 4719 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337516 4719 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337528 4719 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337534 4719 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337539 4719 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337544 4719 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337548 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337555 4719 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337561 4719 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337566 4719 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337572 4719 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337578 4719 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337582 4719 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337587 4719 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337591 4719 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337596 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337601 4719 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337605 4719 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337610 4719 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337615 4719 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337619 4719 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337624 4719 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337628 4719 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337632 4719 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337637 4719 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337642 4719 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337647 4719 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337651 4719 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337656 4719 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337660 4719 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337665 4719 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337671 4719 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337676 4719 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337680 4719 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337684 4719 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337688 4719 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337691 4719 feature_gate.go:330] unrecognized feature gate: Example Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337695 4719 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337698 4719 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337702 4719 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337705 4719 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337709 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337713 4719 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337716 4719 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337720 4719 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337723 4719 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337727 4719 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337730 4719 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337734 4719 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337737 4719 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337741 4719 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337745 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337749 4719 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337752 4719 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337755 4719 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337759 4719 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337763 4719 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337767 4719 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337770 4719 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337775 4719 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337780 4719 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337785 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337789 4719 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337792 4719 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337796 4719 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337800 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337804 4719 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337807 4719 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.337811 4719 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.337819 4719 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338497 4719 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338514 4719 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338520 4719 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338525 4719 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338529 4719 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338533 4719 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338543 4719 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338549 4719 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338554 4719 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338559 4719 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338565 4719 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338571 4719 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338576 4719 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338582 4719 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338588 4719 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338593 4719 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338598 4719 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338603 4719 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338611 4719 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338616 4719 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338622 4719 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338628 4719 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338633 4719 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338638 4719 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338642 4719 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338647 4719 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338652 4719 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338657 4719 feature_gate.go:330] unrecognized feature gate: Example Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338661 4719 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338666 4719 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338673 4719 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338678 4719 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338683 4719 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338687 4719 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338693 4719 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338698 4719 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338703 4719 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338707 4719 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338712 4719 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338716 4719 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338721 4719 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338725 4719 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338730 4719 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338738 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338743 4719 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338749 4719 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338753 4719 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338758 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338763 4719 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338768 4719 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338772 4719 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338778 4719 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338785 4719 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338790 4719 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338795 4719 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338801 4719 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338809 4719 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338815 4719 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338820 4719 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338824 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338829 4719 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338834 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338840 4719 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338846 4719 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338852 4719 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338859 4719 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338881 4719 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338904 4719 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338912 4719 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338917 4719 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.338928 4719 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.338937 4719 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.339421 4719 server.go:940] "Client rotation is on, will bootstrap in background" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.344887 4719 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.345011 4719 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.345657 4719 server.go:997] "Starting client certificate rotation" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.345682 4719 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.346035 4719 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-09 14:00:06.50062671 +0000 UTC Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.346183 4719 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.354196 4719 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.356037 4719 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.356252 4719 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.364388 4719 log.go:25] "Validated CRI v1 runtime API" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.387179 4719 log.go:25] "Validated CRI v1 image API" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.389507 4719 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.393532 4719 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-15-11-28-07-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.393578 4719 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.405298 4719 manager.go:217] Machine: {Timestamp:2025-12-15 12:17:19.403846225 +0000 UTC m=+0.346139275 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199484928 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:7eb9371f-defc-4e63-8dd6-10092c139afd BootID:751f1c28-6acd-4965-8e8f-c2d4a0a91bc0 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076109 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599742464 Type:vfs Inodes:3076109 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:f6:e6:3b Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:f6:e6:3b Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:80:f4:a7 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:12:30:8b Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:23:64:28 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:cb:ce:2d Speed:-1 Mtu:1496} {Name:eth10 MacAddress:26:6e:d9:c8:23:66 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:46:63:62:1a:e3:20 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199484928 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.405539 4719 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.405894 4719 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.406348 4719 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.406529 4719 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.406572 4719 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.407127 4719 topology_manager.go:138] "Creating topology manager with none policy" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.407148 4719 container_manager_linux.go:303] "Creating device plugin manager" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.407333 4719 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.407360 4719 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.407613 4719 state_mem.go:36] "Initialized new in-memory state store" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.408615 4719 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.409904 4719 kubelet.go:418] "Attempting to sync node with API server" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.409928 4719 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.409956 4719 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.409972 4719 kubelet.go:324] "Adding apiserver pod source" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.409988 4719 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.412048 4719 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.412522 4719 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.413273 4719 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.413528 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.413614 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.413638 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.413728 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414558 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414584 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414592 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414599 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414612 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414620 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414628 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414639 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414646 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414654 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414680 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.414735 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.415004 4719 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.415523 4719 server.go:1280] "Started kubelet" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.415689 4719 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.416461 4719 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.416056 4719 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.417498 4719 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:19 crc systemd[1]: Started Kubernetes Kubelet. Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.419500 4719 server.go:460] "Adding debug handlers to kubelet server" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.420957 4719 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.421005 4719 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.421321 4719 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 10:24:12.865098539 +0000 UTC Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.421383 4719 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 550h6m53.443718954s for next certificate rotation Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.422286 4719 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.422302 4719 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.422431 4719 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.423388 4719 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.423723 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="200ms" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.424344 4719 factory.go:55] Registering systemd factory Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.424365 4719 factory.go:221] Registration of the systemd container factory successfully Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.424801 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.424903 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.426418 4719 factory.go:153] Registering CRI-O factory Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.426457 4719 factory.go:221] Registration of the crio container factory successfully Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.426531 4719 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.426569 4719 factory.go:103] Registering Raw factory Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.426593 4719 manager.go:1196] Started watching for new ooms in manager Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.426428 4719 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188162b092602ea9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-15 12:17:19.415488169 +0000 UTC m=+0.357781199,LastTimestamp:2025-12-15 12:17:19.415488169 +0000 UTC m=+0.357781199,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.427258 4719 manager.go:319] Starting recovery of all containers Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440298 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440371 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440389 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440403 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440416 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440430 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440444 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440458 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440477 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440494 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440507 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440519 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440535 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440554 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440570 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440605 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440621 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440636 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440651 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440664 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440692 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440707 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440722 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440739 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440752 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440765 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440782 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440796 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440811 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440826 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440842 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440897 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440914 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440943 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440959 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440972 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.440989 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441006 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441042 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441059 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441076 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441092 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441108 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441125 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441142 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441158 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441172 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441187 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441200 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441212 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441224 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441238 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441255 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441269 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441283 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441296 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441327 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441339 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441351 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441366 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441380 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441396 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441408 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441422 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441434 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441446 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441457 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441470 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441480 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441490 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441500 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441511 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441522 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441534 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441546 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441557 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441567 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441578 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441589 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441599 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441610 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441622 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441632 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441643 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441654 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441665 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441677 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441689 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441700 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441710 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441724 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441735 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441747 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441757 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441771 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441783 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441796 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441809 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441820 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441834 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441846 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441881 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441896 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441910 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441929 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441944 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441958 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441970 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441982 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.441996 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442009 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442021 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442032 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442043 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442055 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442067 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442079 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442091 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442104 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442118 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442130 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442141 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442153 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442165 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442178 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442188 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442200 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442212 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442223 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442234 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442247 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442261 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442274 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442287 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442302 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442315 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442342 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442369 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442386 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442401 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442418 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442434 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442450 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442463 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442477 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442489 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442501 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442513 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442526 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442539 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442554 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442567 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442583 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442597 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442610 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442623 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442635 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442646 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442664 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442676 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442688 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442699 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442713 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442725 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442737 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442750 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442768 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442780 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442791 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442811 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442834 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442924 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442940 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442954 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442968 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442982 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.442995 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443010 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443022 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443034 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443045 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443056 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443067 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443119 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443132 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443144 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443156 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443169 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.443181 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445393 4719 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445460 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445481 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445497 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445515 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445533 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445557 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445575 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445595 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445614 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445630 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445647 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445662 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445678 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445693 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445709 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445724 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445739 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445756 4719 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445770 4719 reconstruct.go:97] "Volume reconstruction finished" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.445781 4719 reconciler.go:26] "Reconciler: start to sync state" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.450665 4719 manager.go:324] Recovery completed Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.463899 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.466489 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.466548 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.466562 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.467331 4719 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.467345 4719 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.467370 4719 state_mem.go:36] "Initialized new in-memory state store" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.472529 4719 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.474713 4719 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.474754 4719 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.474789 4719 kubelet.go:2335] "Starting kubelet main sync loop" Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.474843 4719 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.479193 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.479334 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.479747 4719 policy_none.go:49] "None policy: Start" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.480491 4719 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.480542 4719 state_mem.go:35] "Initializing new in-memory state store" Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.523463 4719 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.552485 4719 manager.go:334] "Starting Device Plugin manager" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.552548 4719 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.552564 4719 server.go:79] "Starting device plugin registration server" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.553000 4719 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.553016 4719 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.553174 4719 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.553283 4719 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.553290 4719 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.559417 4719 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.575837 4719 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.575945 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.576940 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.576972 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.576981 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.577105 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.577529 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.577632 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.577723 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.577763 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.577780 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.577980 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.578478 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.578515 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581126 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581164 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581176 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581530 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581552 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581570 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581704 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581786 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581968 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581747 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.581872 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.582238 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.583096 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.583208 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.583232 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.583514 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.583596 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.583613 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.583660 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.584006 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.584057 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.585246 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.585268 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.585277 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.585422 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.585449 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.585706 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.585791 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.585867 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.586136 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.586169 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.586181 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.624798 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="400ms" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648500 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648571 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648598 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648618 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648643 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648662 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648681 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648719 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648738 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648756 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648775 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648794 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648809 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648825 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.648842 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.653463 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.654968 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.655082 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.655176 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.655287 4719 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.655788 4719 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750325 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750436 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750463 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750482 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750502 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750522 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750554 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750597 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750626 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750612 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750653 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750773 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750818 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750832 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750855 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750904 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750925 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750938 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750952 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750969 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750979 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.751001 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.751033 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.750730 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.751061 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.751101 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.751100 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.751082 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.751162 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.751241 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.856134 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.857547 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.857591 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.857602 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.857638 4719 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 15 12:17:19 crc kubenswrapper[4719]: E1215 12:17:19.858147 4719 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.908951 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.916678 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.942382 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.945467 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-461b64dbe4992971118bda15e8831fa3d9e84ca598dc94d46399f4638135d562 WatchSource:0}: Error finding container 461b64dbe4992971118bda15e8831fa3d9e84ca598dc94d46399f4638135d562: Status 404 returned error can't find the container with id 461b64dbe4992971118bda15e8831fa3d9e84ca598dc94d46399f4638135d562 Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.948582 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-4cd02f122d5007cc36f9018c1eee8ad9742003aad6937829920618b01ae69a06 WatchSource:0}: Error finding container 4cd02f122d5007cc36f9018c1eee8ad9742003aad6937829920618b01ae69a06: Status 404 returned error can't find the container with id 4cd02f122d5007cc36f9018c1eee8ad9742003aad6937829920618b01ae69a06 Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.959052 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: I1215 12:17:19.966091 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 15 12:17:19 crc kubenswrapper[4719]: W1215 12:17:19.989888 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-dfb399a3065042f4be6e2f70c5f9640ca1e97483b95d1826dc00cd0fff76ece9 WatchSource:0}: Error finding container dfb399a3065042f4be6e2f70c5f9640ca1e97483b95d1826dc00cd0fff76ece9: Status 404 returned error can't find the container with id dfb399a3065042f4be6e2f70c5f9640ca1e97483b95d1826dc00cd0fff76ece9 Dec 15 12:17:20 crc kubenswrapper[4719]: E1215 12:17:20.025782 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="800ms" Dec 15 12:17:20 crc kubenswrapper[4719]: W1215 12:17:20.241505 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:20 crc kubenswrapper[4719]: E1215 12:17:20.241640 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.258531 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.260265 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.260301 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.260317 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.260343 4719 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 15 12:17:20 crc kubenswrapper[4719]: E1215 12:17:20.260607 4719 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.418262 4719 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:20 crc kubenswrapper[4719]: W1215 12:17:20.424607 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:20 crc kubenswrapper[4719]: E1215 12:17:20.424705 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Dec 15 12:17:20 crc kubenswrapper[4719]: W1215 12:17:20.463946 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:20 crc kubenswrapper[4719]: E1215 12:17:20.464033 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.484938 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"260b9864653f37af01256617281e26639ac5aaa77ee05b430a02fc5a7c6aec64"} Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.485046 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"461b64dbe4992971118bda15e8831fa3d9e84ca598dc94d46399f4638135d562"} Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.485144 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.486203 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.486228 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.486237 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.490103 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="58eac6ee43de2a094e653292f18561931ad03cc3ada3537c5fe86c234e9228c7" exitCode=0 Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.490255 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"58eac6ee43de2a094e653292f18561931ad03cc3ada3537c5fe86c234e9228c7"} Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.490289 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4cd02f122d5007cc36f9018c1eee8ad9742003aad6937829920618b01ae69a06"} Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.490437 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.492143 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.492166 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.492175 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.494926 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346"} Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.494993 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"76f17dd420c1494e55f729ce7acc8a08bfca84008c369efaf390571802f14946"} Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.495155 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.496179 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.496199 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.496207 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.497580 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054"} Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.497617 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dfb399a3065042f4be6e2f70c5f9640ca1e97483b95d1826dc00cd0fff76ece9"} Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.499214 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650"} Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.499238 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b3540fddc299a6193a0a4875b3ef08476d349322aef7dd8a1f023a21415a3a01"} Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.499305 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.500781 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.500831 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:20 crc kubenswrapper[4719]: I1215 12:17:20.500843 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:20 crc kubenswrapper[4719]: W1215 12:17:20.638991 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:20 crc kubenswrapper[4719]: E1215 12:17:20.639069 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Dec 15 12:17:20 crc kubenswrapper[4719]: E1215 12:17:20.828007 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="1.6s" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.061209 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.062987 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.063033 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.063047 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.063076 4719 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 15 12:17:21 crc kubenswrapper[4719]: E1215 12:17:21.063772 4719 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.419223 4719 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.505083 4719 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650" exitCode=0 Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.505175 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650"} Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.505332 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.506137 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.506165 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.506174 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.508947 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.514147 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.514480 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.514541 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.514797 4719 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="260b9864653f37af01256617281e26639ac5aaa77ee05b430a02fc5a7c6aec64" exitCode=0 Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.514871 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"260b9864653f37af01256617281e26639ac5aaa77ee05b430a02fc5a7c6aec64"} Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.515184 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.516341 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.516368 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.516382 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.519268 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e0ec5c38713b367c92d9d714b3b310cffc7a9722452e4e702cea0e7c5c651b33"} Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.519387 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.520445 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.520466 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.520476 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.522645 4719 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346" exitCode=0 Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.522699 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346"} Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.522717 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725"} Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.522728 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c"} Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.522739 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95"} Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.522818 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.523686 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.523775 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.523847 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.525346 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168"} Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.525374 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5"} Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.525388 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943"} Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.525511 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.526089 4719 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.526316 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.526352 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.526366 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:21 crc kubenswrapper[4719]: E1215 12:17:21.527523 4719 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Dec 15 12:17:21 crc kubenswrapper[4719]: I1215 12:17:21.901908 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.532643 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac"} Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.532721 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2"} Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.532750 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02"} Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.532759 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e"} Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.532773 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092"} Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.533004 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.534132 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.534167 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.534193 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.535415 4719 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e98b0c72c02048d0aaa646ee097dcc29e9a1f0d7b4654428d7f2d28bd3c2c652" exitCode=0 Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.535441 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e98b0c72c02048d0aaa646ee097dcc29e9a1f0d7b4654428d7f2d28bd3c2c652"} Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.535608 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.535625 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.536768 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.536790 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.536827 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.536839 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.536803 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.536942 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.654572 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.664891 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.666097 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.666122 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.666133 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:22 crc kubenswrapper[4719]: I1215 12:17:22.666164 4719 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.540091 4719 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b0c3035819ad1892b3b8e376d155dcda2427ceb864feef972da53b2866537339" exitCode=0 Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.540205 4719 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.540240 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.540750 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b0c3035819ad1892b3b8e376d155dcda2427ceb864feef972da53b2866537339"} Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.540838 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.541196 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.541621 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.541650 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.541658 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.542004 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.542048 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.542062 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.542180 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.542215 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.542230 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:23 crc kubenswrapper[4719]: I1215 12:17:23.553212 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.547477 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ef55f4727c7c7827314c7852e94810138bf3170faee2cf628c9aa20d7a8aa444"} Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.547558 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"efe5a1b418ce4815a63a5395c3b85d42b958d937696657ba65a584bc1e76f0ae"} Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.547578 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6a1930760367ab0745557ad25571252376c042563fdb3c510f13ba9580c4d869"} Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.547592 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8c3c234548bd0641e81d72bdd4beab790aa1c5f94eddfbf23fb2cdb6c9295da9"} Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.547603 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dc82de3e459599aae8762fe616cf272d5a1332e6bff0166fb1be7b5108e14765"} Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.547602 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.547734 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.548741 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.548780 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.548790 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.548807 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.548841 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:24 crc kubenswrapper[4719]: I1215 12:17:24.548850 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:25 crc kubenswrapper[4719]: I1215 12:17:25.828501 4719 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.145390 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.145614 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.146836 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.146891 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.146903 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.356385 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.356546 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.358006 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.358068 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.358081 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.361642 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.552553 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.552680 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.554098 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.554166 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:26 crc kubenswrapper[4719]: I1215 12:17:26.554180 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.083217 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.083519 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.084800 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.084842 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.084873 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.555785 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.557189 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.557238 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.557250 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.636267 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.636464 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.637733 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.637768 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:27 crc kubenswrapper[4719]: I1215 12:17:27.637779 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:28 crc kubenswrapper[4719]: I1215 12:17:28.599497 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:28 crc kubenswrapper[4719]: I1215 12:17:28.599761 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:28 crc kubenswrapper[4719]: I1215 12:17:28.600844 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:28 crc kubenswrapper[4719]: I1215 12:17:28.600896 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:28 crc kubenswrapper[4719]: I1215 12:17:28.600907 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:29 crc kubenswrapper[4719]: I1215 12:17:29.150690 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 15 12:17:29 crc kubenswrapper[4719]: I1215 12:17:29.150974 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:29 crc kubenswrapper[4719]: I1215 12:17:29.152117 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:29 crc kubenswrapper[4719]: I1215 12:17:29.152153 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:29 crc kubenswrapper[4719]: I1215 12:17:29.152164 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:29 crc kubenswrapper[4719]: E1215 12:17:29.559721 4719 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 15 12:17:30 crc kubenswrapper[4719]: I1215 12:17:30.585724 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:30 crc kubenswrapper[4719]: I1215 12:17:30.586673 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:30 crc kubenswrapper[4719]: I1215 12:17:30.589281 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:30 crc kubenswrapper[4719]: I1215 12:17:30.589328 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:30 crc kubenswrapper[4719]: I1215 12:17:30.589348 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:31 crc kubenswrapper[4719]: I1215 12:17:31.600619 4719 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 15 12:17:31 crc kubenswrapper[4719]: I1215 12:17:31.600749 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 15 12:17:32 crc kubenswrapper[4719]: I1215 12:17:32.283782 4719 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 15 12:17:32 crc kubenswrapper[4719]: I1215 12:17:32.283892 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 15 12:17:32 crc kubenswrapper[4719]: I1215 12:17:32.420060 4719 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 15 12:17:32 crc kubenswrapper[4719]: E1215 12:17:32.431355 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 15 12:17:32 crc kubenswrapper[4719]: E1215 12:17:32.666882 4719 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 15 12:17:32 crc kubenswrapper[4719]: W1215 12:17:32.869117 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 15 12:17:32 crc kubenswrapper[4719]: I1215 12:17:32.869239 4719 trace.go:236] Trace[1726377461]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (15-Dec-2025 12:17:22.867) (total time: 10001ms): Dec 15 12:17:32 crc kubenswrapper[4719]: Trace[1726377461]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (12:17:32.869) Dec 15 12:17:32 crc kubenswrapper[4719]: Trace[1726377461]: [10.001898455s] [10.001898455s] END Dec 15 12:17:32 crc kubenswrapper[4719]: E1215 12:17:32.869274 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 15 12:17:32 crc kubenswrapper[4719]: W1215 12:17:32.994228 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 15 12:17:32 crc kubenswrapper[4719]: I1215 12:17:32.994367 4719 trace.go:236] Trace[1630098394]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (15-Dec-2025 12:17:22.993) (total time: 10001ms): Dec 15 12:17:32 crc kubenswrapper[4719]: Trace[1630098394]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (12:17:32.994) Dec 15 12:17:32 crc kubenswrapper[4719]: Trace[1630098394]: [10.001234077s] [10.001234077s] END Dec 15 12:17:32 crc kubenswrapper[4719]: E1215 12:17:32.994401 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 15 12:17:33 crc kubenswrapper[4719]: W1215 12:17:33.250957 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 15 12:17:33 crc kubenswrapper[4719]: I1215 12:17:33.251095 4719 trace.go:236] Trace[1399955708]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (15-Dec-2025 12:17:23.249) (total time: 10001ms): Dec 15 12:17:33 crc kubenswrapper[4719]: Trace[1399955708]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (12:17:33.250) Dec 15 12:17:33 crc kubenswrapper[4719]: Trace[1399955708]: [10.001511995s] [10.001511995s] END Dec 15 12:17:33 crc kubenswrapper[4719]: E1215 12:17:33.251133 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 15 12:17:33 crc kubenswrapper[4719]: W1215 12:17:33.559146 4719 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 15 12:17:33 crc kubenswrapper[4719]: I1215 12:17:33.559282 4719 trace.go:236] Trace[1594070989]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (15-Dec-2025 12:17:23.558) (total time: 10001ms): Dec 15 12:17:33 crc kubenswrapper[4719]: Trace[1594070989]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (12:17:33.559) Dec 15 12:17:33 crc kubenswrapper[4719]: Trace[1594070989]: [10.001125353s] [10.001125353s] END Dec 15 12:17:33 crc kubenswrapper[4719]: E1215 12:17:33.559318 4719 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 15 12:17:33 crc kubenswrapper[4719]: I1215 12:17:33.616282 4719 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 15 12:17:33 crc kubenswrapper[4719]: I1215 12:17:33.616375 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 15 12:17:33 crc kubenswrapper[4719]: I1215 12:17:33.621537 4719 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 15 12:17:33 crc kubenswrapper[4719]: I1215 12:17:33.621637 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 15 12:17:35 crc kubenswrapper[4719]: I1215 12:17:35.867592 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:35 crc kubenswrapper[4719]: I1215 12:17:35.869352 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:35 crc kubenswrapper[4719]: I1215 12:17:35.869394 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:35 crc kubenswrapper[4719]: I1215 12:17:35.869405 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:35 crc kubenswrapper[4719]: I1215 12:17:35.869432 4719 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 15 12:17:35 crc kubenswrapper[4719]: E1215 12:17:35.875164 4719 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 15 12:17:36 crc kubenswrapper[4719]: I1215 12:17:36.170622 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 15 12:17:36 crc kubenswrapper[4719]: I1215 12:17:36.170845 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:36 crc kubenswrapper[4719]: I1215 12:17:36.171958 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:36 crc kubenswrapper[4719]: I1215 12:17:36.171987 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:36 crc kubenswrapper[4719]: I1215 12:17:36.171996 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:36 crc kubenswrapper[4719]: I1215 12:17:36.187899 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 15 12:17:36 crc kubenswrapper[4719]: I1215 12:17:36.581934 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:36 crc kubenswrapper[4719]: I1215 12:17:36.582922 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:36 crc kubenswrapper[4719]: I1215 12:17:36.582967 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:36 crc kubenswrapper[4719]: I1215 12:17:36.582984 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.089022 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.089240 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.090517 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.090554 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.090564 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.093260 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.338663 4719 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.585573 4719 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.585635 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.586901 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.586936 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:37 crc kubenswrapper[4719]: I1215 12:17:37.586947 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.144661 4719 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.421180 4719 apiserver.go:52] "Watching apiserver" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.426625 4719 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.427022 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.427551 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.427618 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.427690 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.427737 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.427906 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.427943 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.427991 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.427996 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.429095 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.431035 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.431045 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.431187 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.431036 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.431764 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.432166 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.432881 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.433188 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.433340 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.470479 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.491396 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.508551 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.523876 4719 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.524356 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.533733 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.541906 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.553724 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.568324 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.579609 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.597694 4719 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.631922 4719 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.664829 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.666664 4719 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46498->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.666726 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46498->192.168.126.11:17697: read: connection reset by peer" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.667028 4719 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.667068 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.672269 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.683515 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.692560 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698521 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698578 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698610 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698640 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698676 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698707 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698738 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698769 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698797 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698822 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698875 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698940 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.698973 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699000 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699025 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699054 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699081 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699110 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699139 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699168 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699199 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699228 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699260 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699289 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699319 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699353 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699389 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699420 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699453 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699485 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699540 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699583 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699617 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699651 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699689 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699732 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699770 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699813 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699881 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699922 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699956 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.699988 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700021 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700053 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700093 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700126 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700158 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700191 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700237 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700269 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700302 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700340 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700382 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700415 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700447 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700482 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700517 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700552 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700584 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700612 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700642 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700673 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700702 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700729 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700757 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700784 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700812 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700844 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700909 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700943 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.700974 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701001 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701028 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701055 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701083 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701111 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701141 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701168 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701199 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701234 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701254 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701270 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701303 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701340 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701376 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701427 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701461 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701500 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701530 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701566 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701601 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701631 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701636 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701704 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701743 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701781 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701818 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701878 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701914 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701946 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.701978 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702010 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702045 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702084 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702122 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702155 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702188 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702223 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702257 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702287 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702320 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702359 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702396 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702433 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702467 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702502 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702539 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702603 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702638 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702670 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702709 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702744 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702778 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702813 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702848 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702930 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702964 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702993 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703027 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703056 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703087 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703132 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703169 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703201 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703234 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703271 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703307 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703338 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703368 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703398 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703434 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703466 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703504 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703536 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703571 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703603 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703633 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703667 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703706 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703745 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703782 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703818 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703877 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703914 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703946 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703979 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704014 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704050 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704087 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704123 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704160 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704195 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704231 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704269 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704308 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704346 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704463 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704502 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704538 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704576 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704614 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704653 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704693 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704735 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704773 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704811 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704853 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705278 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705303 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705328 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705355 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705381 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705405 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705432 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705458 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705483 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705509 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705535 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705560 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705584 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705608 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705633 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705659 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705693 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705729 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705772 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705813 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705907 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705952 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705999 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706039 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706082 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706125 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706167 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706213 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706297 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706336 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706375 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706416 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706454 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706494 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706574 4719 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706600 4719 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.707026 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702045 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702369 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.702684 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703713 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.703926 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704128 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704706 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704835 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.704938 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705011 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705025 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705033 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705173 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705248 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705256 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705299 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705438 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705455 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705466 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705687 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.705874 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706022 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706078 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706115 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706166 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706183 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706335 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706483 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706511 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706681 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706713 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706767 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.706868 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.707126 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.707290 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.707311 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.707561 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.707577 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.707786 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.708122 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.708424 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.708434 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.709519 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.708445 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.708240 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.708928 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.709000 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.709100 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.710283 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.710347 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.710644 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.710534 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.711100 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.711378 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.711643 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.712933 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.712997 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.713256 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.713252 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.713319 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.713521 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.713810 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.713822 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.714458 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.714577 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.714743 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.714759 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.714758 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.715107 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.715197 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.715364 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.715597 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.715673 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.715846 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.715959 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.717661 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.717729 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.718086 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.718322 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.718477 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.718685 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.718699 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.719121 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.719306 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.719307 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.719674 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.719962 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.720316 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.720925 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.721167 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.721495 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.722586 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.722807 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.722980 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.723603 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.723706 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.723905 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.722078 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.723982 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.724206 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.724402 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.724415 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.724531 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.724552 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.724551 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.724564 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.724592 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.724924 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.725116 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.725404 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.726214 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.726337 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:17:39.226318828 +0000 UTC m=+20.168611858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.726417 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.726459 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.727270 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.726498 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.726551 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.726582 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.726592 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.726658 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.726827 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.727036 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.727069 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.727242 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.727285 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.727444 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.727647 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.727886 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.727991 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.728111 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.728379 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.728481 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.728684 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.728937 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.729121 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.729180 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.729949 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.729969 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.730018 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.730046 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.730247 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.730938 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.731256 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.731564 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.731601 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.731618 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.731739 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732314 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732483 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732531 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732616 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732676 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732697 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732739 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732760 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732803 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732831 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.732895 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.733353 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.733516 4719 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.733702 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:39.23366071 +0000 UTC m=+20.175953830 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.733706 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.733882 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.734063 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.734084 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.734266 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.734598 4719 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.734648 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:39.234636571 +0000 UTC m=+20.176929671 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.734793 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.736972 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.743166 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.743329 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.743764 4719 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.744732 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.749390 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.750615 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.751265 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.751343 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.756789 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.757070 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.757420 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.757708 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.757723 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.757734 4719 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.757781 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:39.257766813 +0000 UTC m=+20.200059843 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.758144 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.758607 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.758619 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.758627 4719 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.758654 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:39.258644971 +0000 UTC m=+20.200938001 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.769879 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.770034 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.770040 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.770064 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.770175 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.770197 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.770447 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.770555 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.771798 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.771943 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.772091 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.772276 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.772525 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.772925 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.773996 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.775024 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.775077 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.774196 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.776099 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.777072 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.783554 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.784071 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.784193 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.784338 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.789295 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.794213 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: W1215 12:17:38.798529 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-728d0c7b0da6b70afaab545b652ec0648cf5fe7bccee7373031d5286df651c92 WatchSource:0}: Error finding container 728d0c7b0da6b70afaab545b652ec0648cf5fe7bccee7373031d5286df651c92: Status 404 returned error can't find the container with id 728d0c7b0da6b70afaab545b652ec0648cf5fe7bccee7373031d5286df651c92 Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.802403 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.803033 4719 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 15 12:17:38 crc kubenswrapper[4719]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Dec 15 12:17:38 crc kubenswrapper[4719]: if [[ -f "/env/_master" ]]; then Dec 15 12:17:38 crc kubenswrapper[4719]: set -o allexport Dec 15 12:17:38 crc kubenswrapper[4719]: source "/env/_master" Dec 15 12:17:38 crc kubenswrapper[4719]: set +o allexport Dec 15 12:17:38 crc kubenswrapper[4719]: fi Dec 15 12:17:38 crc kubenswrapper[4719]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Dec 15 12:17:38 crc kubenswrapper[4719]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Dec 15 12:17:38 crc kubenswrapper[4719]: ho_enable="--enable-hybrid-overlay" Dec 15 12:17:38 crc kubenswrapper[4719]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Dec 15 12:17:38 crc kubenswrapper[4719]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Dec 15 12:17:38 crc kubenswrapper[4719]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Dec 15 12:17:38 crc kubenswrapper[4719]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Dec 15 12:17:38 crc kubenswrapper[4719]: --webhook-cert-dir="/etc/webhook-cert" \ Dec 15 12:17:38 crc kubenswrapper[4719]: --webhook-host=127.0.0.1 \ Dec 15 12:17:38 crc kubenswrapper[4719]: --webhook-port=9743 \ Dec 15 12:17:38 crc kubenswrapper[4719]: ${ho_enable} \ Dec 15 12:17:38 crc kubenswrapper[4719]: --enable-interconnect \ Dec 15 12:17:38 crc kubenswrapper[4719]: --disable-approver \ Dec 15 12:17:38 crc kubenswrapper[4719]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Dec 15 12:17:38 crc kubenswrapper[4719]: --wait-for-kubernetes-api=200s \ Dec 15 12:17:38 crc kubenswrapper[4719]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Dec 15 12:17:38 crc kubenswrapper[4719]: --loglevel="${LOGLEVEL}" Dec 15 12:17:38 crc kubenswrapper[4719]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 15 12:17:38 crc kubenswrapper[4719]: > logger="UnhandledError" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.804325 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.805016 4719 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 15 12:17:38 crc kubenswrapper[4719]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Dec 15 12:17:38 crc kubenswrapper[4719]: if [[ -f "/env/_master" ]]; then Dec 15 12:17:38 crc kubenswrapper[4719]: set -o allexport Dec 15 12:17:38 crc kubenswrapper[4719]: source "/env/_master" Dec 15 12:17:38 crc kubenswrapper[4719]: set +o allexport Dec 15 12:17:38 crc kubenswrapper[4719]: fi Dec 15 12:17:38 crc kubenswrapper[4719]: Dec 15 12:17:38 crc kubenswrapper[4719]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Dec 15 12:17:38 crc kubenswrapper[4719]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Dec 15 12:17:38 crc kubenswrapper[4719]: --disable-webhook \ Dec 15 12:17:38 crc kubenswrapper[4719]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Dec 15 12:17:38 crc kubenswrapper[4719]: --loglevel="${LOGLEVEL}" Dec 15 12:17:38 crc kubenswrapper[4719]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 15 12:17:38 crc kubenswrapper[4719]: > logger="UnhandledError" Dec 15 12:17:38 crc kubenswrapper[4719]: E1215 12:17:38.806185 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807287 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807330 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807397 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807408 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807417 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807428 4719 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807438 4719 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807447 4719 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807455 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807463 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807472 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807481 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807489 4719 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807498 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807507 4719 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807516 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807524 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807533 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807542 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807551 4719 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807561 4719 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807572 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807581 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807592 4719 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807600 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807608 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807617 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807626 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807634 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807642 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807650 4719 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807658 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807666 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807674 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807683 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807691 4719 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807699 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807708 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807716 4719 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807725 4719 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807735 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807744 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807753 4719 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807761 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807770 4719 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807778 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807788 4719 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807798 4719 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807807 4719 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807816 4719 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807825 4719 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807835 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807844 4719 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807852 4719 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807861 4719 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807888 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807896 4719 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807905 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807915 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807923 4719 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807932 4719 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807940 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807949 4719 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807957 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807980 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807988 4719 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.807997 4719 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808006 4719 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808015 4719 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808032 4719 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808042 4719 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808052 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808061 4719 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808069 4719 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808078 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808086 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808095 4719 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808104 4719 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808113 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808125 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808133 4719 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808142 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808150 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808159 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808167 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808175 4719 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808184 4719 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808192 4719 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808200 4719 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808222 4719 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808230 4719 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808239 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808247 4719 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808256 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808265 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808274 4719 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808283 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808292 4719 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808300 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808320 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808328 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808337 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808345 4719 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808353 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808362 4719 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808369 4719 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808377 4719 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808397 4719 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808406 4719 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808414 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808422 4719 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808431 4719 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808440 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808449 4719 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808457 4719 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808465 4719 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808494 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808512 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808527 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808536 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808545 4719 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808554 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808562 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808571 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808581 4719 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808589 4719 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808597 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808605 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808613 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808621 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808629 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808638 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808646 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808654 4719 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808662 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808671 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808674 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808708 4719 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808758 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808776 4719 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808791 4719 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808805 4719 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808836 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808849 4719 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808907 4719 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808920 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808933 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808966 4719 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808979 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.808991 4719 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809002 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809013 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809025 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809055 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809067 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809078 4719 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809090 4719 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809144 4719 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809156 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809168 4719 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809180 4719 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809265 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809281 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809294 4719 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809339 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809354 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809369 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809413 4719 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809428 4719 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809441 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809454 4719 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809485 4719 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809497 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809510 4719 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809610 4719 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809624 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809672 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809685 4719 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809845 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809918 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809931 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809946 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809958 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.809991 4719 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.810003 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.810017 4719 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.810030 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.810041 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.810075 4719 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.810090 4719 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.810103 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.810117 4719 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.810164 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.810179 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.814109 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.815848 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.818327 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.822474 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.833380 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.842504 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.851268 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.861044 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.910957 4719 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:38 crc kubenswrapper[4719]: I1215 12:17:38.910988 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.056943 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 15 12:17:39 crc kubenswrapper[4719]: W1215 12:17:39.069619 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-e6c331342252e57e3e8e9a623b4c351b85b1ed3cbf7271749c25933c4a86cda1 WatchSource:0}: Error finding container e6c331342252e57e3e8e9a623b4c351b85b1ed3cbf7271749c25933c4a86cda1: Status 404 returned error can't find the container with id e6c331342252e57e3e8e9a623b4c351b85b1ed3cbf7271749c25933c4a86cda1 Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.071805 4719 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 15 12:17:39 crc kubenswrapper[4719]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Dec 15 12:17:39 crc kubenswrapper[4719]: set -o allexport Dec 15 12:17:39 crc kubenswrapper[4719]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Dec 15 12:17:39 crc kubenswrapper[4719]: source /etc/kubernetes/apiserver-url.env Dec 15 12:17:39 crc kubenswrapper[4719]: else Dec 15 12:17:39 crc kubenswrapper[4719]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Dec 15 12:17:39 crc kubenswrapper[4719]: exit 1 Dec 15 12:17:39 crc kubenswrapper[4719]: fi Dec 15 12:17:39 crc kubenswrapper[4719]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Dec 15 12:17:39 crc kubenswrapper[4719]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 15 12:17:39 crc kubenswrapper[4719]: > logger="UnhandledError" Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.072897 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.074476 4719 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.090784 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 15 12:17:39 crc kubenswrapper[4719]: W1215 12:17:39.104614 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-a9d15b7e832d088092cd6c6820fe2e2512114c560a9c2d31bc2e1e0e00d9bde6 WatchSource:0}: Error finding container a9d15b7e832d088092cd6c6820fe2e2512114c560a9c2d31bc2e1e0e00d9bde6: Status 404 returned error can't find the container with id a9d15b7e832d088092cd6c6820fe2e2512114c560a9c2d31bc2e1e0e00d9bde6 Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.312624 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.312717 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.312743 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.312760 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:17:40.312742199 +0000 UTC m=+21.255035229 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.312784 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.312830 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.312840 4719 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.312895 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:40.312886673 +0000 UTC m=+21.255179703 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.312934 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.312947 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.312958 4719 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.312981 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.313020 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.313026 4719 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.313041 4719 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.312988 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:40.312980716 +0000 UTC m=+21.255273746 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.313060 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:40.313054299 +0000 UTC m=+21.255347329 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.313110 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:40.313090379 +0000 UTC m=+21.255383409 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.480313 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.481050 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.482756 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.483696 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.485052 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.485772 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.486610 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.487944 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.488750 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.490575 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.491813 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.492057 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.493531 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.494354 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.495271 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.496099 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.496848 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.497645 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.498224 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.499036 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.500937 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.501586 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.502408 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.503045 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.503414 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.503972 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.504502 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.505312 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.506182 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.506800 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.509249 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.510214 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.511178 4719 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.511376 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.514702 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.515472 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.516756 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.518301 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.520980 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.523177 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.524493 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.525829 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.526453 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.527306 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.528311 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.529592 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.530936 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.533744 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.534886 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.535697 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.537120 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.538105 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.538337 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.539642 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.540300 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.541136 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.542038 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.542898 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.543564 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.546343 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.557061 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.591554 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"728d0c7b0da6b70afaab545b652ec0648cf5fe7bccee7373031d5286df651c92"} Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.595222 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.599913 4719 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac" exitCode=255 Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.600097 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac"} Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.602037 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a9d15b7e832d088092cd6c6820fe2e2512114c560a9c2d31bc2e1e0e00d9bde6"} Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.603742 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e6c331342252e57e3e8e9a623b4c351b85b1ed3cbf7271749c25933c4a86cda1"} Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.605460 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: E1215 12:17:39.608896 4719 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.609443 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.610621 4719 scope.go:117] "RemoveContainer" containerID="f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.619280 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.624150 4719 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.630155 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.638238 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.647823 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.658813 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.667427 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.676387 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.688126 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.701262 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.711439 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.728273 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.746078 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.764067 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:39 crc kubenswrapper[4719]: I1215 12:17:39.784309 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.324192 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.324487 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:17:42.324444092 +0000 UTC m=+23.266737122 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.324747 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.324788 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.324816 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.324841 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.324979 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.324998 4719 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.325073 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.325116 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.325132 4719 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.325135 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:42.325108134 +0000 UTC m=+23.267401184 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.324989 4719 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.325019 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.325213 4719 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.325195 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:42.325174376 +0000 UTC m=+23.267467566 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.325243 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:42.325233468 +0000 UTC m=+23.267526688 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.325263 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:42.325252478 +0000 UTC m=+23.267545728 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.475777 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.475925 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.475961 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.476008 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.476102 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:40 crc kubenswrapper[4719]: E1215 12:17:40.476253 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.607873 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.609540 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97"} Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.609841 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.610942 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065"} Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.612597 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3"} Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.612626 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f"} Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.626703 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.641003 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.661918 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.676441 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.702603 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.724088 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.740842 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.755852 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.772785 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.786625 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.805350 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.821821 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.835929 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.850188 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.864883 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:40 crc kubenswrapper[4719]: I1215 12:17:40.885648 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.276212 4719 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.282479 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.282592 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.282619 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.282733 4719 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.291835 4719 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.292232 4719 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.293995 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.294145 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.294219 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.294294 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.294379 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.312770 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.317535 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.317573 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.317585 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.317600 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.317613 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.329190 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.334088 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.334567 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.334774 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.334968 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.335133 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.339696 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.339837 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.339891 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.339920 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.339960 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:17:46.339928991 +0000 UTC m=+27.282222021 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.340021 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340064 4719 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340108 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340133 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:46.340116957 +0000 UTC m=+27.282409987 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340152 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340168 4719 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340177 4719 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340206 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:46.34019976 +0000 UTC m=+27.282492790 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340206 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340259 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340275 4719 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340285 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:46.340259682 +0000 UTC m=+27.282552852 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.340343 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:46.340316654 +0000 UTC m=+27.282609844 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.355662 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.359998 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.360024 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.360032 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.360045 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.360055 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.383254 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.389641 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.389670 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.389679 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.389691 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.389701 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.414590 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.414728 4719 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.416677 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.416745 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.416757 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.416778 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.416797 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.475840 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.475900 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.475901 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.476010 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.476092 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:42 crc kubenswrapper[4719]: E1215 12:17:42.476171 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.518755 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.518794 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.518803 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.518818 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.518827 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.619919 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63"} Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.620237 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.620273 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.620283 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.620295 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.620304 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.633604 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.647807 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.661090 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.674986 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.688684 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.701420 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.715205 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.724521 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.724589 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.724603 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.724620 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.724636 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.733720 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.827395 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.827438 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.827449 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.827464 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.827478 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.930443 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.930492 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.930504 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.930518 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.930527 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:42Z","lastTransitionTime":"2025-12-15T12:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.934791 4719 csr.go:261] certificate signing request csr-5q4dr is approved, waiting to be issued Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.962327 4719 csr.go:257] certificate signing request csr-5q4dr is issued Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.988278 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-w4tb7"] Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.988658 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-w4tb7" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.995113 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 15 12:17:42 crc kubenswrapper[4719]: I1215 12:17:42.996034 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.002235 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.020738 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.032732 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.032786 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.032800 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.032817 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.032830 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:43Z","lastTransitionTime":"2025-12-15T12:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.042275 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.073420 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.109627 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.131637 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.135713 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.135747 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.135758 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.135773 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.135785 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:43Z","lastTransitionTime":"2025-12-15T12:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.146454 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d246m\" (UniqueName: \"kubernetes.io/projected/02119208-83cb-48f3-b72e-e31a01753830-kube-api-access-d246m\") pod \"node-resolver-w4tb7\" (UID: \"02119208-83cb-48f3-b72e-e31a01753830\") " pod="openshift-dns/node-resolver-w4tb7" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.146522 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/02119208-83cb-48f3-b72e-e31a01753830-hosts-file\") pod \"node-resolver-w4tb7\" (UID: \"02119208-83cb-48f3-b72e-e31a01753830\") " pod="openshift-dns/node-resolver-w4tb7" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.150751 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.171477 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.207978 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.238080 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.238121 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.238132 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.238148 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.238158 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:43Z","lastTransitionTime":"2025-12-15T12:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.247668 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d246m\" (UniqueName: \"kubernetes.io/projected/02119208-83cb-48f3-b72e-e31a01753830-kube-api-access-d246m\") pod \"node-resolver-w4tb7\" (UID: \"02119208-83cb-48f3-b72e-e31a01753830\") " pod="openshift-dns/node-resolver-w4tb7" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.247727 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/02119208-83cb-48f3-b72e-e31a01753830-hosts-file\") pod \"node-resolver-w4tb7\" (UID: \"02119208-83cb-48f3-b72e-e31a01753830\") " pod="openshift-dns/node-resolver-w4tb7" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.247811 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/02119208-83cb-48f3-b72e-e31a01753830-hosts-file\") pod \"node-resolver-w4tb7\" (UID: \"02119208-83cb-48f3-b72e-e31a01753830\") " pod="openshift-dns/node-resolver-w4tb7" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.257695 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.292718 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d246m\" (UniqueName: \"kubernetes.io/projected/02119208-83cb-48f3-b72e-e31a01753830-kube-api-access-d246m\") pod \"node-resolver-w4tb7\" (UID: \"02119208-83cb-48f3-b72e-e31a01753830\") " pod="openshift-dns/node-resolver-w4tb7" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.301586 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-w4tb7" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.344223 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.344268 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.344281 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.344300 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.344312 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:43Z","lastTransitionTime":"2025-12-15T12:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.446884 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.446925 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.446935 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.446951 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.446961 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:43Z","lastTransitionTime":"2025-12-15T12:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.504885 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-dwfzn"] Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.505765 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.524458 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-ff9fm"] Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.524840 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.525030 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.525032 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.525384 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.525382 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.526555 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-lcgn4"] Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.526748 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.528113 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.528709 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.529069 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.529714 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.530292 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.530344 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.532237 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.532238 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.551142 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.551177 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.551188 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.551204 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.551216 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:43Z","lastTransitionTime":"2025-12-15T12:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.553215 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.574833 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.594641 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.621691 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.624346 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-w4tb7" event={"ID":"02119208-83cb-48f3-b72e-e31a01753830","Type":"ContainerStarted","Data":"f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.624398 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-w4tb7" event={"ID":"02119208-83cb-48f3-b72e-e31a01753830","Type":"ContainerStarted","Data":"f2bbee68bd0f0c5a5e44269abb14cc73e0eb8d8b167ffc56f3fc0a8642893984"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.637352 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653494 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653534 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653544 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653559 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653569 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:43Z","lastTransitionTime":"2025-12-15T12:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653661 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/70777f65-d38a-4352-8e39-a8681d7996f2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653709 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-system-cni-dir\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653731 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-cnibin\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653754 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eaba47bc-655f-4987-b3c6-b8d49ad5c128-mcd-auth-proxy-config\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653772 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkqnp\" (UniqueName: \"kubernetes.io/projected/eaba47bc-655f-4987-b3c6-b8d49ad5c128-kube-api-access-rkqnp\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653789 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlt9t\" (UniqueName: \"kubernetes.io/projected/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-kube-api-access-mlt9t\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653864 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-cni-dir\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653888 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-etc-kubernetes\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653906 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-os-release\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653928 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-system-cni-dir\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.653979 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/70777f65-d38a-4352-8e39-a8681d7996f2-cni-binary-copy\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654020 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-run-netns\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654078 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-var-lib-kubelet\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654103 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/eaba47bc-655f-4987-b3c6-b8d49ad5c128-rootfs\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654120 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-var-lib-cni-multus\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654145 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-run-k8s-cni-cncf-io\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654183 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsd2g\" (UniqueName: \"kubernetes.io/projected/70777f65-d38a-4352-8e39-a8681d7996f2-kube-api-access-bsd2g\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654220 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-cni-binary-copy\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654260 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-daemon-config\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654294 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-os-release\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654326 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654360 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-conf-dir\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654394 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-run-multus-certs\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654419 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/eaba47bc-655f-4987-b3c6-b8d49ad5c128-proxy-tls\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654436 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-socket-dir-parent\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654458 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-var-lib-cni-bin\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654475 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-hostroot\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.654500 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-cnibin\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.656202 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.671072 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.687678 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.707623 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.736702 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.755611 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-os-release\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.755774 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-os-release\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.755953 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-system-cni-dir\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756121 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/70777f65-d38a-4352-8e39-a8681d7996f2-cni-binary-copy\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756146 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-run-netns\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756179 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-var-lib-kubelet\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756207 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/eaba47bc-655f-4987-b3c6-b8d49ad5c128-rootfs\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756241 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-run-k8s-cni-cncf-io\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756227 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-run-netns\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756257 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-var-lib-cni-multus\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756285 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-var-lib-kubelet\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756307 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsd2g\" (UniqueName: \"kubernetes.io/projected/70777f65-d38a-4352-8e39-a8681d7996f2-kube-api-access-bsd2g\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756312 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-run-k8s-cni-cncf-io\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756421 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/eaba47bc-655f-4987-b3c6-b8d49ad5c128-rootfs\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756427 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756464 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756476 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756441 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-var-lib-cni-multus\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756440 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-cni-binary-copy\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756493 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756568 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:43Z","lastTransitionTime":"2025-12-15T12:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756613 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-daemon-config\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756653 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-os-release\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756690 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756704 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-os-release\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756723 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-conf-dir\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756759 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-run-multus-certs\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756795 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/eaba47bc-655f-4987-b3c6-b8d49ad5c128-proxy-tls\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756823 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-socket-dir-parent\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756825 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-conf-dir\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756847 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-var-lib-cni-bin\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756900 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-hostroot\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756968 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-cnibin\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.756988 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/70777f65-d38a-4352-8e39-a8681d7996f2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757027 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-system-cni-dir\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757048 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-cnibin\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757079 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eaba47bc-655f-4987-b3c6-b8d49ad5c128-mcd-auth-proxy-config\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757098 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkqnp\" (UniqueName: \"kubernetes.io/projected/eaba47bc-655f-4987-b3c6-b8d49ad5c128-kube-api-access-rkqnp\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757126 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-cni-dir\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757144 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-etc-kubernetes\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757166 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlt9t\" (UniqueName: \"kubernetes.io/projected/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-kube-api-access-mlt9t\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757369 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-cni-binary-copy\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757390 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-daemon-config\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757576 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-system-cni-dir\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757655 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-socket-dir-parent\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757671 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-etc-kubernetes\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757712 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-var-lib-cni-bin\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757750 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-host-run-multus-certs\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757776 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-system-cni-dir\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757755 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-hostroot\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757838 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-cnibin\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757849 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-cnibin\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757933 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eaba47bc-655f-4987-b3c6-b8d49ad5c128-mcd-auth-proxy-config\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.757635 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-multus-cni-dir\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.758630 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/70777f65-d38a-4352-8e39-a8681d7996f2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.758806 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/70777f65-d38a-4352-8e39-a8681d7996f2-cni-binary-copy\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.764261 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/70777f65-d38a-4352-8e39-a8681d7996f2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.765529 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/eaba47bc-655f-4987-b3c6-b8d49ad5c128-proxy-tls\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.768391 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.789459 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkqnp\" (UniqueName: \"kubernetes.io/projected/eaba47bc-655f-4987-b3c6-b8d49ad5c128-kube-api-access-rkqnp\") pod \"machine-config-daemon-ff9fm\" (UID: \"eaba47bc-655f-4987-b3c6-b8d49ad5c128\") " pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.790320 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlt9t\" (UniqueName: \"kubernetes.io/projected/e63b3458-45ea-47c9-a9ff-7651b7dd4cca-kube-api-access-mlt9t\") pod \"multus-lcgn4\" (UID: \"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\") " pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.791240 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsd2g\" (UniqueName: \"kubernetes.io/projected/70777f65-d38a-4352-8e39-a8681d7996f2-kube-api-access-bsd2g\") pod \"multus-additional-cni-plugins-dwfzn\" (UID: \"70777f65-d38a-4352-8e39-a8681d7996f2\") " pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.819825 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.822962 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: W1215 12:17:43.836245 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70777f65_d38a_4352_8e39_a8681d7996f2.slice/crio-9df8338be73f376e7f6ccb5238a70bd5521ab2e57a9cd351a1df85a90dbc9a47 WatchSource:0}: Error finding container 9df8338be73f376e7f6ccb5238a70bd5521ab2e57a9cd351a1df85a90dbc9a47: Status 404 returned error can't find the container with id 9df8338be73f376e7f6ccb5238a70bd5521ab2e57a9cd351a1df85a90dbc9a47 Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.837909 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.848548 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lcgn4" Dec 15 12:17:43 crc kubenswrapper[4719]: W1215 12:17:43.852796 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeaba47bc_655f_4987_b3c6_b8d49ad5c128.slice/crio-9d518fea7b041df2bcdefe2551c769f9300d9a5be822a611ec296bef7246f379 WatchSource:0}: Error finding container 9d518fea7b041df2bcdefe2551c769f9300d9a5be822a611ec296bef7246f379: Status 404 returned error can't find the container with id 9d518fea7b041df2bcdefe2551c769f9300d9a5be822a611ec296bef7246f379 Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.860646 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.860704 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.860717 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.860736 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.860748 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:43Z","lastTransitionTime":"2025-12-15T12:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.889345 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.919409 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.954536 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-j8shp"] Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.955338 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.960354 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.960549 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.960595 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.960558 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.960793 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.960933 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.960949 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.961146 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.963780 4719 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-15 12:12:42 +0000 UTC, rotation deadline is 2026-09-07 19:31:43.097223193 +0000 UTC Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.963799 4719 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6391h13m59.133425973s for next certificate rotation Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.970745 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.970768 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.970776 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.970788 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.970797 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:43Z","lastTransitionTime":"2025-12-15T12:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:43 crc kubenswrapper[4719]: I1215 12:17:43.996231 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:43Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.024368 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.036650 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.055804 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062553 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-config\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062586 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znshh\" (UniqueName: \"kubernetes.io/projected/d1afde33-4ec0-4b20-9915-1722f99997fe-kube-api-access-znshh\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062605 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-ovn-kubernetes\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062631 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062688 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-systemd\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062708 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-script-lib\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062733 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-slash\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062757 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-netd\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062771 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-env-overrides\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062788 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1afde33-4ec0-4b20-9915-1722f99997fe-ovn-node-metrics-cert\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062805 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-log-socket\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062827 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-var-lib-openvswitch\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062844 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-etc-openvswitch\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062895 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-node-log\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062920 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-netns\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062942 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-ovn\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062960 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-kubelet\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062975 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-openvswitch\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.062992 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-systemd-units\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.063008 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-bin\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.073764 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.074193 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.074216 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.074226 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.074242 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.074253 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:44Z","lastTransitionTime":"2025-12-15T12:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.090033 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.108034 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.121459 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.141741 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.158820 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163618 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-netns\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163663 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-ovn\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163685 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-kubelet\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163701 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-openvswitch\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163717 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-bin\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163733 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-systemd-units\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163747 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-netns\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163771 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-config\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163784 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-openvswitch\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163802 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-systemd-units\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163883 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-bin\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.163900 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-kubelet\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164030 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-ovn\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164481 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-config\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164532 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znshh\" (UniqueName: \"kubernetes.io/projected/d1afde33-4ec0-4b20-9915-1722f99997fe-kube-api-access-znshh\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164567 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-ovn-kubernetes\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164592 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164639 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-ovn-kubernetes\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164667 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-systemd\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164686 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-script-lib\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164702 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-slash\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165246 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-netd\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165267 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-env-overrides\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165686 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-log-socket\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165708 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1afde33-4ec0-4b20-9915-1722f99997fe-ovn-node-metrics-cert\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165746 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-log-socket\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164731 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-systemd\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164749 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165647 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-env-overrides\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.164841 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-slash\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165204 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-script-lib\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165302 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-netd\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165909 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-var-lib-openvswitch\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165932 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-var-lib-openvswitch\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165951 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-etc-openvswitch\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165971 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-etc-openvswitch\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.165994 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-node-log\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.166027 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-node-log\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.169589 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1afde33-4ec0-4b20-9915-1722f99997fe-ovn-node-metrics-cert\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.177560 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.177601 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.177613 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.177631 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.177643 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:44Z","lastTransitionTime":"2025-12-15T12:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.179853 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znshh\" (UniqueName: \"kubernetes.io/projected/d1afde33-4ec0-4b20-9915-1722f99997fe-kube-api-access-znshh\") pod \"ovnkube-node-j8shp\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.182329 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.202072 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.221432 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.237506 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.250065 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.271638 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.274967 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.289444 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.289481 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.289494 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.289515 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.289526 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:44Z","lastTransitionTime":"2025-12-15T12:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.315566 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.341799 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.368036 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.392368 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.392400 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.392408 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.392422 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.392430 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:44Z","lastTransitionTime":"2025-12-15T12:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.397445 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.475286 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:44 crc kubenswrapper[4719]: E1215 12:17:44.475441 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.475303 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:44 crc kubenswrapper[4719]: E1215 12:17:44.475524 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.475300 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:44 crc kubenswrapper[4719]: E1215 12:17:44.475571 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.495639 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.495687 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.495696 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.495710 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.495721 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:44Z","lastTransitionTime":"2025-12-15T12:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.602413 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.602470 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.602481 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.602507 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.602518 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:44Z","lastTransitionTime":"2025-12-15T12:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.628429 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5" exitCode=0 Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.628514 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.628572 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"9f3b314ffd9a827cbb7b0808c98f3c094e1b82cac64eb7e739dada30ad41321b"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.631429 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lcgn4" event={"ID":"e63b3458-45ea-47c9-a9ff-7651b7dd4cca","Type":"ContainerStarted","Data":"01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.631471 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lcgn4" event={"ID":"e63b3458-45ea-47c9-a9ff-7651b7dd4cca","Type":"ContainerStarted","Data":"4f6188cc21f4a6a3ddc6a0511f7083a3cd1e8bd603851175260f148ebce0ead6"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.648692 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.648697 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.648749 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.648763 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"9d518fea7b041df2bcdefe2551c769f9300d9a5be822a611ec296bef7246f379"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.651830 4719 generic.go:334] "Generic (PLEG): container finished" podID="70777f65-d38a-4352-8e39-a8681d7996f2" containerID="32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa" exitCode=0 Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.651891 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" event={"ID":"70777f65-d38a-4352-8e39-a8681d7996f2","Type":"ContainerDied","Data":"32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.651918 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" event={"ID":"70777f65-d38a-4352-8e39-a8681d7996f2","Type":"ContainerStarted","Data":"9df8338be73f376e7f6ccb5238a70bd5521ab2e57a9cd351a1df85a90dbc9a47"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.662774 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.679704 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.693268 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.705743 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.705783 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.705792 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.705807 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.705819 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:44Z","lastTransitionTime":"2025-12-15T12:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.709707 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.723114 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.738324 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.751021 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.762772 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.781280 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.804099 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.808978 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.809026 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.809035 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.809048 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.809057 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:44Z","lastTransitionTime":"2025-12-15T12:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.817935 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.831762 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.845266 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.860636 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.873504 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.912337 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.912381 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.912394 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.912410 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.912422 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:44Z","lastTransitionTime":"2025-12-15T12:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:44 crc kubenswrapper[4719]: I1215 12:17:44.975809 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.004671 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.014918 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.014965 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.014977 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.015003 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.015017 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:45Z","lastTransitionTime":"2025-12-15T12:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.024113 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.044951 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.058360 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.076474 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.089118 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.105242 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.116342 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.118348 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.118393 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.118408 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.118425 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.118436 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:45Z","lastTransitionTime":"2025-12-15T12:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.134069 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.220161 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.220187 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.220197 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.220214 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.220224 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:45Z","lastTransitionTime":"2025-12-15T12:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.324119 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.324568 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.324578 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.324593 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.324602 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:45Z","lastTransitionTime":"2025-12-15T12:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.427711 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.427762 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.427774 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.427791 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.427803 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:45Z","lastTransitionTime":"2025-12-15T12:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.530357 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.530384 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.530395 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.530408 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.530419 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:45Z","lastTransitionTime":"2025-12-15T12:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.632410 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.632920 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.632929 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.632943 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.632953 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:45Z","lastTransitionTime":"2025-12-15T12:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.658749 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.658805 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.658819 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.658832 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.658843 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.661664 4719 generic.go:334] "Generic (PLEG): container finished" podID="70777f65-d38a-4352-8e39-a8681d7996f2" containerID="01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a" exitCode=0 Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.661695 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" event={"ID":"70777f65-d38a-4352-8e39-a8681d7996f2","Type":"ContainerDied","Data":"01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.676290 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.691986 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.708455 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.723423 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.735254 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.735300 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.735312 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.735332 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.735344 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:45Z","lastTransitionTime":"2025-12-15T12:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.737789 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.751274 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.771566 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.788931 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.803880 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.823777 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.835372 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.837093 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.837116 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.837125 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.837141 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.837150 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:45Z","lastTransitionTime":"2025-12-15T12:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.847202 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.862344 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.881815 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zbphz"] Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.882318 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.884112 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.885089 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.885343 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.885523 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.905354 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.917782 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.930612 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.939451 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.939483 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.939491 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.939505 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.939514 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:45Z","lastTransitionTime":"2025-12-15T12:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.942416 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.953788 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.962692 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.975837 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.984300 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86c27adb-05f9-4c0b-96b3-eb144b6f1c8a-host\") pod \"node-ca-zbphz\" (UID: \"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\") " pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.984346 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/86c27adb-05f9-4c0b-96b3-eb144b6f1c8a-serviceca\") pod \"node-ca-zbphz\" (UID: \"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\") " pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.984365 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqgnp\" (UniqueName: \"kubernetes.io/projected/86c27adb-05f9-4c0b-96b3-eb144b6f1c8a-kube-api-access-jqgnp\") pod \"node-ca-zbphz\" (UID: \"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\") " pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.987918 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:45 crc kubenswrapper[4719]: I1215 12:17:45.998452 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:45Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.010528 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.019579 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.031248 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.042406 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.042445 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.042454 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.042468 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.042478 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:46Z","lastTransitionTime":"2025-12-15T12:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.044090 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.055394 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.085987 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86c27adb-05f9-4c0b-96b3-eb144b6f1c8a-host\") pod \"node-ca-zbphz\" (UID: \"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\") " pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.086032 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/86c27adb-05f9-4c0b-96b3-eb144b6f1c8a-serviceca\") pod \"node-ca-zbphz\" (UID: \"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\") " pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.086055 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqgnp\" (UniqueName: \"kubernetes.io/projected/86c27adb-05f9-4c0b-96b3-eb144b6f1c8a-kube-api-access-jqgnp\") pod \"node-ca-zbphz\" (UID: \"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\") " pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.086083 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86c27adb-05f9-4c0b-96b3-eb144b6f1c8a-host\") pod \"node-ca-zbphz\" (UID: \"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\") " pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.087274 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/86c27adb-05f9-4c0b-96b3-eb144b6f1c8a-serviceca\") pod \"node-ca-zbphz\" (UID: \"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\") " pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.106615 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqgnp\" (UniqueName: \"kubernetes.io/projected/86c27adb-05f9-4c0b-96b3-eb144b6f1c8a-kube-api-access-jqgnp\") pod \"node-ca-zbphz\" (UID: \"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\") " pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.145947 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.145999 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.146016 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.146036 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.146050 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:46Z","lastTransitionTime":"2025-12-15T12:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.198552 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zbphz" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.250010 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.250036 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.250045 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.250060 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.250070 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:46Z","lastTransitionTime":"2025-12-15T12:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.352807 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.352841 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.352850 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.352890 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.352900 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:46Z","lastTransitionTime":"2025-12-15T12:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.389113 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.389255 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:17:54.389228039 +0000 UTC m=+35.331521069 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.389292 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.389333 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.389358 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.389390 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.389514 4719 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.389577 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:54.38955895 +0000 UTC m=+35.331852030 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.389968 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.389995 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.390009 4719 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.390051 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:54.390041615 +0000 UTC m=+35.332334655 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.390066 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.390116 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.390131 4719 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.390202 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:54.39018934 +0000 UTC m=+35.332482370 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.390244 4719 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.390269 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:17:54.390262652 +0000 UTC m=+35.332555682 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.455852 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.455912 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.455924 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.455940 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.455951 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:46Z","lastTransitionTime":"2025-12-15T12:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.475606 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.475718 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.475781 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.475732 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.475896 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:46 crc kubenswrapper[4719]: E1215 12:17:46.475937 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.559080 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.559120 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.559130 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.559146 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.559158 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:46Z","lastTransitionTime":"2025-12-15T12:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.661651 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.661693 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.661705 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.661723 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.661735 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:46Z","lastTransitionTime":"2025-12-15T12:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.667071 4719 generic.go:334] "Generic (PLEG): container finished" podID="70777f65-d38a-4352-8e39-a8681d7996f2" containerID="dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962" exitCode=0 Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.667155 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" event={"ID":"70777f65-d38a-4352-8e39-a8681d7996f2","Type":"ContainerDied","Data":"dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.668735 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zbphz" event={"ID":"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a","Type":"ContainerStarted","Data":"1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.668758 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zbphz" event={"ID":"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a","Type":"ContainerStarted","Data":"e80f99f22e22cbf9b173b4c21de46f9c58ecd9b63b9a7cdd2ea3d57df32d5512"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.673206 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.685104 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.696422 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.711217 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.725321 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.736984 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.749600 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.762023 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.764793 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.764839 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.764853 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.764893 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.764906 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:46Z","lastTransitionTime":"2025-12-15T12:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.775102 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.787105 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.796932 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.809638 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.828280 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.845758 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.859629 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.867465 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.867503 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.867512 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.867526 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.867535 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:46Z","lastTransitionTime":"2025-12-15T12:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.873580 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.895602 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.909571 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.924223 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.955464 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.970048 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.970089 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.970098 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.970113 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.970124 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:46Z","lastTransitionTime":"2025-12-15T12:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:46 crc kubenswrapper[4719]: I1215 12:17:46.978198 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:46Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.004583 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.018575 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.031219 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.042764 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.055954 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.066732 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.072060 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.072106 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.072117 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.072134 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.072149 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:47Z","lastTransitionTime":"2025-12-15T12:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.082889 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.096685 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.174748 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.174802 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.174812 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.174826 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.174836 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:47Z","lastTransitionTime":"2025-12-15T12:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.277774 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.277839 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.277855 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.277916 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.277929 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:47Z","lastTransitionTime":"2025-12-15T12:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.382412 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.382484 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.382505 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.382533 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.382551 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:47Z","lastTransitionTime":"2025-12-15T12:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.485270 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.485354 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.485372 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.485396 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.485414 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:47Z","lastTransitionTime":"2025-12-15T12:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.588165 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.588263 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.588289 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.588321 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.588345 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:47Z","lastTransitionTime":"2025-12-15T12:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.681409 4719 generic.go:334] "Generic (PLEG): container finished" podID="70777f65-d38a-4352-8e39-a8681d7996f2" containerID="4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8" exitCode=0 Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.681470 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" event={"ID":"70777f65-d38a-4352-8e39-a8681d7996f2","Type":"ContainerDied","Data":"4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8"} Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.691005 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.691076 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.691092 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.691113 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.691127 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:47Z","lastTransitionTime":"2025-12-15T12:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.704531 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.720705 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.746423 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.762360 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.780753 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.794008 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.795208 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.795242 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.795251 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.795266 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.795283 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:47Z","lastTransitionTime":"2025-12-15T12:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.819054 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.840917 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.852790 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.868292 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.889968 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.897194 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.897230 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.897243 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.897260 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.897273 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:47Z","lastTransitionTime":"2025-12-15T12:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.906092 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.921086 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:47 crc kubenswrapper[4719]: I1215 12:17:47.933717 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:47Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.001572 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.001615 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.001625 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.001640 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.001650 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:48Z","lastTransitionTime":"2025-12-15T12:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.103828 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.103901 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.103919 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.103946 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.103963 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:48Z","lastTransitionTime":"2025-12-15T12:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.207018 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.207082 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.207098 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.207123 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.207139 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:48Z","lastTransitionTime":"2025-12-15T12:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.310532 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.310591 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.310610 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.310633 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.310653 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:48Z","lastTransitionTime":"2025-12-15T12:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.413963 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.414012 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.414022 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.414041 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.414054 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:48Z","lastTransitionTime":"2025-12-15T12:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.475667 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.475706 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.475772 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:48 crc kubenswrapper[4719]: E1215 12:17:48.475885 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:48 crc kubenswrapper[4719]: E1215 12:17:48.476245 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:48 crc kubenswrapper[4719]: E1215 12:17:48.476338 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.516675 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.516726 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.516736 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.516757 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.516769 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:48Z","lastTransitionTime":"2025-12-15T12:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.619394 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.619441 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.619451 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.619472 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.619484 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:48Z","lastTransitionTime":"2025-12-15T12:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.694757 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" event={"ID":"70777f65-d38a-4352-8e39-a8681d7996f2","Type":"ContainerStarted","Data":"1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.699757 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.709978 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.722202 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.722263 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.722273 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.722293 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.722496 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:48Z","lastTransitionTime":"2025-12-15T12:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.728284 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.748977 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.763593 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.779247 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.792710 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.811111 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.825492 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.825802 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.825879 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.825894 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.825922 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.825939 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:48Z","lastTransitionTime":"2025-12-15T12:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.846336 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.860472 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.876544 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.890176 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.903722 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.919406 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:48Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.928890 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.928949 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.928963 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.928990 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:48 crc kubenswrapper[4719]: I1215 12:17:48.929004 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:48Z","lastTransitionTime":"2025-12-15T12:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.031726 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.031798 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.031813 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.031854 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.031890 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:49Z","lastTransitionTime":"2025-12-15T12:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.135355 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.135387 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.135395 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.135408 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.135418 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:49Z","lastTransitionTime":"2025-12-15T12:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.237872 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.238208 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.238217 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.238235 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.238245 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:49Z","lastTransitionTime":"2025-12-15T12:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.340926 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.340990 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.341011 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.341038 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.341057 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:49Z","lastTransitionTime":"2025-12-15T12:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.347036 4719 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.443647 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.443691 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.443702 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.443720 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.443731 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:49Z","lastTransitionTime":"2025-12-15T12:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.492674 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.505730 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.518897 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.536548 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.546309 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.546354 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.546368 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.546389 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.546403 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:49Z","lastTransitionTime":"2025-12-15T12:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.552421 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.571047 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.588046 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.607476 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.622686 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.639365 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.653267 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.653324 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.653338 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.653358 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.653371 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:49Z","lastTransitionTime":"2025-12-15T12:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.658652 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.673100 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.687675 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.707983 4719 generic.go:334] "Generic (PLEG): container finished" podID="70777f65-d38a-4352-8e39-a8681d7996f2" containerID="1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9" exitCode=0 Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.708065 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" event={"ID":"70777f65-d38a-4352-8e39-a8681d7996f2","Type":"ContainerDied","Data":"1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.708789 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.725023 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.750818 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.757020 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.757077 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.757090 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.757115 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.757130 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:49Z","lastTransitionTime":"2025-12-15T12:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.767588 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.783500 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.802525 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.817376 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.833906 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.846900 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.861030 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.861069 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.861080 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.861104 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.861114 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:49Z","lastTransitionTime":"2025-12-15T12:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.862160 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.877238 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.895147 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.910883 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.927424 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.943397 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:49Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.963368 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.963422 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.963436 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.963491 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:49 crc kubenswrapper[4719]: I1215 12:17:49.963505 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:49Z","lastTransitionTime":"2025-12-15T12:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.066498 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.066546 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.066559 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.066576 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.066590 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:50Z","lastTransitionTime":"2025-12-15T12:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.169621 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.169692 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.169723 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.169749 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.169778 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:50Z","lastTransitionTime":"2025-12-15T12:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.272177 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.272236 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.272254 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.272276 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.272296 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:50Z","lastTransitionTime":"2025-12-15T12:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.374939 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.374987 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.375004 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.375027 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.375045 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:50Z","lastTransitionTime":"2025-12-15T12:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.475944 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.476022 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.476087 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:50 crc kubenswrapper[4719]: E1215 12:17:50.476181 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:50 crc kubenswrapper[4719]: E1215 12:17:50.476311 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:50 crc kubenswrapper[4719]: E1215 12:17:50.476567 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.478063 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.478099 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.478108 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.478119 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.478132 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:50Z","lastTransitionTime":"2025-12-15T12:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.581036 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.581080 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.581088 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.581103 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.581113 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:50Z","lastTransitionTime":"2025-12-15T12:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.683894 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.683943 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.683955 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.683971 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.683984 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:50Z","lastTransitionTime":"2025-12-15T12:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.715075 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" event={"ID":"70777f65-d38a-4352-8e39-a8681d7996f2","Type":"ContainerStarted","Data":"d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.786425 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.786480 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.786500 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.786524 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.786541 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:50Z","lastTransitionTime":"2025-12-15T12:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.888491 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.888533 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.888542 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.888554 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.888563 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:50Z","lastTransitionTime":"2025-12-15T12:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.991996 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.992019 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.992027 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.992042 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:50 crc kubenswrapper[4719]: I1215 12:17:50.992051 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:50Z","lastTransitionTime":"2025-12-15T12:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.095492 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.095537 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.095549 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.095568 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.095579 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:51Z","lastTransitionTime":"2025-12-15T12:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.198480 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.198519 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.198559 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.198575 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.198584 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:51Z","lastTransitionTime":"2025-12-15T12:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.300838 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.300894 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.300905 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.300922 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.300931 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:51Z","lastTransitionTime":"2025-12-15T12:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.404321 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.404354 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.404365 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.404382 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.404393 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:51Z","lastTransitionTime":"2025-12-15T12:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.507323 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.507364 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.507378 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.507392 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.507404 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:51Z","lastTransitionTime":"2025-12-15T12:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.610850 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.610906 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.610916 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.610931 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.610942 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:51Z","lastTransitionTime":"2025-12-15T12:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.713776 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.714275 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.714287 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.714302 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.714313 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:51Z","lastTransitionTime":"2025-12-15T12:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.720919 4719 generic.go:334] "Generic (PLEG): container finished" podID="70777f65-d38a-4352-8e39-a8681d7996f2" containerID="d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256" exitCode=0 Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.720962 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" event={"ID":"70777f65-d38a-4352-8e39-a8681d7996f2","Type":"ContainerDied","Data":"d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.726312 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.726573 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.739997 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.759908 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.762257 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.782081 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.796195 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.811348 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.816566 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.816598 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.816607 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.816625 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.816636 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:51Z","lastTransitionTime":"2025-12-15T12:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.824112 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.840929 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.854946 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.869034 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.881090 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.892361 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.905525 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.917227 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.920648 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.920706 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.920718 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.920735 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.920747 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:51Z","lastTransitionTime":"2025-12-15T12:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.933102 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.946028 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.962069 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.979450 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:51 crc kubenswrapper[4719]: I1215 12:17:51.992529 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:51Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.009657 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.024355 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.024405 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.024414 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.024440 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.024453 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.026668 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.040971 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.053397 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.064974 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.080978 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.093657 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.106728 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.127579 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.127644 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.127709 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.127761 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.127781 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.130281 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.151495 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.231669 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.231728 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.231740 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.231762 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.231780 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.341096 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.341133 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.341144 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.341162 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.341175 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.444012 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.444066 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.444077 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.444095 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.444107 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.475578 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.475589 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:52 crc kubenswrapper[4719]: E1215 12:17:52.475720 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.475608 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:52 crc kubenswrapper[4719]: E1215 12:17:52.475811 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:52 crc kubenswrapper[4719]: E1215 12:17:52.475885 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.547476 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.547529 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.547541 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.547557 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.547566 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.650204 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.650242 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.650251 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.650455 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.650472 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.729312 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.729364 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.729377 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.729396 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.729695 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.734238 4719 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.734964 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" event={"ID":"70777f65-d38a-4352-8e39-a8681d7996f2","Type":"ContainerStarted","Data":"89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.735172 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.750310 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: E1215 12:17:52.750452 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.757286 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.757320 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.757328 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.757363 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.757374 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.763905 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.764113 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: E1215 12:17:52.769926 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.773354 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.773392 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.773401 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.773416 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.773428 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.779314 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.793944 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: E1215 12:17:52.794324 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.798498 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.798530 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.798539 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.798554 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.798564 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.805869 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: E1215 12:17:52.814160 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.818002 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.823766 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.823813 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.823826 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.823844 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.823876 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.828622 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: E1215 12:17:52.834932 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: E1215 12:17:52.835043 4719 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.836416 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.836439 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.836450 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.836466 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.836479 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.843232 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.864004 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.879450 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.905393 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.920150 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.935315 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.944830 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.944868 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.944877 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.944891 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.944901 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:52Z","lastTransitionTime":"2025-12-15T12:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.949448 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.962607 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.976827 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:52 crc kubenswrapper[4719]: I1215 12:17:52.992461 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:52Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.007134 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.022869 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.037841 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.048181 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.048233 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.048248 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.048268 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.048279 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:53Z","lastTransitionTime":"2025-12-15T12:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.054648 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.065717 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.077774 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.091002 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.100897 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.111952 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.124766 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.144751 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.150758 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.150822 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.150836 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.150869 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.150881 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:53Z","lastTransitionTime":"2025-12-15T12:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.253465 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.253506 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.253515 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.253529 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.253539 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:53Z","lastTransitionTime":"2025-12-15T12:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.356971 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.357015 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.357026 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.357042 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.357051 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:53Z","lastTransitionTime":"2025-12-15T12:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.459916 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.459957 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.459969 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.459984 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.459995 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:53Z","lastTransitionTime":"2025-12-15T12:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.557808 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.561940 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.561973 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.561983 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.561997 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.562008 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:53Z","lastTransitionTime":"2025-12-15T12:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.572536 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.591484 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.604497 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.620317 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.634743 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.652881 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.664893 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.665207 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.665302 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.665413 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.665523 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:53Z","lastTransitionTime":"2025-12-15T12:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.667072 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.683685 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.704110 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.729162 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.738136 4719 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.746548 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.760139 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.768252 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.768290 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.768300 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.768317 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.768328 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:53Z","lastTransitionTime":"2025-12-15T12:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.779507 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.791909 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:53Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.871330 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.871363 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.871374 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.871386 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.871395 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:53Z","lastTransitionTime":"2025-12-15T12:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.974196 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.974681 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.974690 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.974706 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:53 crc kubenswrapper[4719]: I1215 12:17:53.974715 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:53Z","lastTransitionTime":"2025-12-15T12:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.077738 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.077774 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.077786 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.077802 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.077815 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:54Z","lastTransitionTime":"2025-12-15T12:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.179912 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.179960 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.179973 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.179992 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.180011 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:54Z","lastTransitionTime":"2025-12-15T12:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.282090 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.282134 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.282144 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.282157 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.282166 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:54Z","lastTransitionTime":"2025-12-15T12:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.384567 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.384609 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.384621 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.384636 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.384647 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:54Z","lastTransitionTime":"2025-12-15T12:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.475108 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.475172 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.475133 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.475255 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.475302 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.475367 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.479360 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.479406 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.479426 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.479451 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.479478 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479582 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479600 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479605 4719 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479614 4719 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479624 4719 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479652 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-15 12:18:10.479637204 +0000 UTC m=+51.421930234 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479677 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:18:10.479670065 +0000 UTC m=+51.421963095 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479692 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:18:10.479686556 +0000 UTC m=+51.421979586 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479691 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479754 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479782 4719 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479711 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:18:10.479701296 +0000 UTC m=+51.421994326 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:17:54 crc kubenswrapper[4719]: E1215 12:17:54.479931 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:18:10.479898753 +0000 UTC m=+51.422191823 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.487336 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.487358 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.487366 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.487377 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.487385 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:54Z","lastTransitionTime":"2025-12-15T12:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.589587 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.589626 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.589637 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.589654 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.589663 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:54Z","lastTransitionTime":"2025-12-15T12:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.691972 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.692008 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.692018 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.692033 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.692043 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:54Z","lastTransitionTime":"2025-12-15T12:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.740166 4719 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.795094 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.795130 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.795139 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.795158 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.795168 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:54Z","lastTransitionTime":"2025-12-15T12:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.897951 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.897979 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.897988 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.898001 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:54 crc kubenswrapper[4719]: I1215 12:17:54.898010 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:54Z","lastTransitionTime":"2025-12-15T12:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.000273 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.000315 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.000326 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.000344 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.000356 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:55Z","lastTransitionTime":"2025-12-15T12:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.103311 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.103370 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.103394 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.103438 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.103456 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:55Z","lastTransitionTime":"2025-12-15T12:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.211782 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.211832 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.211842 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.211870 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.211883 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:55Z","lastTransitionTime":"2025-12-15T12:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.314638 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.314688 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.314700 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.314719 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.314732 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:55Z","lastTransitionTime":"2025-12-15T12:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.417633 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.417968 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.418097 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.418195 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.418277 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:55Z","lastTransitionTime":"2025-12-15T12:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.521631 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.521682 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.521696 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.521716 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.521729 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:55Z","lastTransitionTime":"2025-12-15T12:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.625298 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.625357 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.625374 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.625400 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.625422 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:55Z","lastTransitionTime":"2025-12-15T12:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.729023 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.729072 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.729084 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.729102 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.729116 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:55Z","lastTransitionTime":"2025-12-15T12:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.745361 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/0.log" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.748739 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8" exitCode=1 Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.748795 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.749662 4719 scope.go:117] "RemoveContainer" containerID="f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.773823 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.793014 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.812768 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.827416 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.831589 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.831622 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.831631 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.831643 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.831652 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:55Z","lastTransitionTime":"2025-12-15T12:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.840062 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.852777 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.867731 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.879799 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.894577 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.908513 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.922071 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.934697 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.934739 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.934752 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.934770 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.934783 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:55Z","lastTransitionTime":"2025-12-15T12:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.942342 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:54Z\\\",\\\"message\\\":\\\"mers/factory.go:160\\\\nI1215 12:17:54.179259 5941 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179277 5941 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1215 12:17:54.179294 5941 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:54.179302 5941 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1215 12:17:54.179328 5941 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1215 12:17:54.179332 5941 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1215 12:17:54.179348 5941 factory.go:656] Stopping watch factory\\\\nI1215 12:17:54.179363 5941 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1215 12:17:54.179372 5941 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1215 12:17:54.179378 5941 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1215 12:17:54.179428 5941 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179767 5941 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1215 12:17:54.180338 5941 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.954717 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:55 crc kubenswrapper[4719]: I1215 12:17:55.969405 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:55Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.037722 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.037773 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.037786 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.037805 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.037818 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:56Z","lastTransitionTime":"2025-12-15T12:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.140629 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.140670 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.140679 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.140693 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.140703 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:56Z","lastTransitionTime":"2025-12-15T12:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.242973 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.243043 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.243063 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.243087 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.243104 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:56Z","lastTransitionTime":"2025-12-15T12:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.345067 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.345112 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.345122 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.345137 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.345146 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:56Z","lastTransitionTime":"2025-12-15T12:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.447506 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.447534 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.447542 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.447554 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.447564 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:56Z","lastTransitionTime":"2025-12-15T12:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.475378 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:56 crc kubenswrapper[4719]: E1215 12:17:56.475516 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.475728 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.475800 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:56 crc kubenswrapper[4719]: E1215 12:17:56.475925 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:56 crc kubenswrapper[4719]: E1215 12:17:56.476016 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.549526 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.549575 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.549587 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.549606 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.549617 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:56Z","lastTransitionTime":"2025-12-15T12:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.652322 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.652357 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.652370 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.652384 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.652392 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:56Z","lastTransitionTime":"2025-12-15T12:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.754143 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.754190 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.754204 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.754220 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.754235 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:56Z","lastTransitionTime":"2025-12-15T12:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.754253 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/0.log" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.759485 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.759643 4719 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.774281 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.784159 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.798668 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.811380 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.824678 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.825725 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9"] Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.826157 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.828550 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.829151 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.841018 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.856244 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.856787 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.856995 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.857007 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.857046 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.857059 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:56Z","lastTransitionTime":"2025-12-15T12:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.877439 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.893501 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.898271 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjdp2\" (UniqueName: \"kubernetes.io/projected/3ae283eb-510a-4419-be95-520ea068a831-kube-api-access-cjdp2\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.898328 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3ae283eb-510a-4419-be95-520ea068a831-env-overrides\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.898363 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3ae283eb-510a-4419-be95-520ea068a831-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.898381 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3ae283eb-510a-4419-be95-520ea068a831-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.905511 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.928557 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:54Z\\\",\\\"message\\\":\\\"mers/factory.go:160\\\\nI1215 12:17:54.179259 5941 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179277 5941 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1215 12:17:54.179294 5941 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:54.179302 5941 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1215 12:17:54.179328 5941 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1215 12:17:54.179332 5941 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1215 12:17:54.179348 5941 factory.go:656] Stopping watch factory\\\\nI1215 12:17:54.179363 5941 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1215 12:17:54.179372 5941 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1215 12:17:54.179378 5941 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1215 12:17:54.179428 5941 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179767 5941 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1215 12:17:54.180338 5941 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.941842 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.954437 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.958953 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.958994 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.959007 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.959022 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.959033 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:56Z","lastTransitionTime":"2025-12-15T12:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.973553 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.986180 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.999544 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjdp2\" (UniqueName: \"kubernetes.io/projected/3ae283eb-510a-4419-be95-520ea068a831-kube-api-access-cjdp2\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.999642 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3ae283eb-510a-4419-be95-520ea068a831-env-overrides\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.999555 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:56Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.999718 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3ae283eb-510a-4419-be95-520ea068a831-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:56 crc kubenswrapper[4719]: I1215 12:17:56.999763 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3ae283eb-510a-4419-be95-520ea068a831-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.000344 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3ae283eb-510a-4419-be95-520ea068a831-env-overrides\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.000640 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3ae283eb-510a-4419-be95-520ea068a831-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.008404 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3ae283eb-510a-4419-be95-520ea068a831-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.016231 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjdp2\" (UniqueName: \"kubernetes.io/projected/3ae283eb-510a-4419-be95-520ea068a831-kube-api-access-cjdp2\") pod \"ovnkube-control-plane-749d76644c-66nd9\" (UID: \"3ae283eb-510a-4419-be95-520ea068a831\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.022687 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:54Z\\\",\\\"message\\\":\\\"mers/factory.go:160\\\\nI1215 12:17:54.179259 5941 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179277 5941 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1215 12:17:54.179294 5941 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:54.179302 5941 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1215 12:17:54.179328 5941 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1215 12:17:54.179332 5941 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1215 12:17:54.179348 5941 factory.go:656] Stopping watch factory\\\\nI1215 12:17:54.179363 5941 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1215 12:17:54.179372 5941 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1215 12:17:54.179378 5941 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1215 12:17:54.179428 5941 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179767 5941 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1215 12:17:54.180338 5941 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.035165 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.049765 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.061547 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.061603 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.061619 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.061639 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.061654 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:57Z","lastTransitionTime":"2025-12-15T12:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.062568 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.077973 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.091997 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.110348 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.126002 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.140139 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.142409 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: W1215 12:17:57.154992 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ae283eb_510a_4419_be95_520ea068a831.slice/crio-52db26b3efbb78fbcdfef022d5b63205a40ea166bfd7c0134cb468d93f9023bc WatchSource:0}: Error finding container 52db26b3efbb78fbcdfef022d5b63205a40ea166bfd7c0134cb468d93f9023bc: Status 404 returned error can't find the container with id 52db26b3efbb78fbcdfef022d5b63205a40ea166bfd7c0134cb468d93f9023bc Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.156797 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.165215 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.165255 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.165265 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.165281 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.165290 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:57Z","lastTransitionTime":"2025-12-15T12:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.171006 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.194912 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.207997 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.268000 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.268028 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.268038 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.268054 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.268067 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:57Z","lastTransitionTime":"2025-12-15T12:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.370788 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.371306 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.371317 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.371332 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.371342 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:57Z","lastTransitionTime":"2025-12-15T12:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.474410 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.474460 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.474471 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.474486 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.474498 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:57Z","lastTransitionTime":"2025-12-15T12:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.576886 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.576922 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.576931 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.576944 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.576954 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:57Z","lastTransitionTime":"2025-12-15T12:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.680227 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.680261 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.680269 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.680283 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.680293 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:57Z","lastTransitionTime":"2025-12-15T12:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.763900 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" event={"ID":"3ae283eb-510a-4419-be95-520ea068a831","Type":"ContainerStarted","Data":"8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.763953 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" event={"ID":"3ae283eb-510a-4419-be95-520ea068a831","Type":"ContainerStarted","Data":"9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.763968 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" event={"ID":"3ae283eb-510a-4419-be95-520ea068a831","Type":"ContainerStarted","Data":"52db26b3efbb78fbcdfef022d5b63205a40ea166bfd7c0134cb468d93f9023bc"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.765425 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/1.log" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.765966 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/0.log" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.768655 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9" exitCode=1 Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.768701 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.768750 4719 scope.go:117] "RemoveContainer" containerID="f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.769494 4719 scope.go:117] "RemoveContainer" containerID="a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9" Dec 15 12:17:57 crc kubenswrapper[4719]: E1215 12:17:57.769668 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.778968 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.783346 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.783640 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.783726 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.783807 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.783975 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:57Z","lastTransitionTime":"2025-12-15T12:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.796957 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.816214 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:54Z\\\",\\\"message\\\":\\\"mers/factory.go:160\\\\nI1215 12:17:54.179259 5941 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179277 5941 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1215 12:17:54.179294 5941 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:54.179302 5941 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1215 12:17:54.179328 5941 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1215 12:17:54.179332 5941 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1215 12:17:54.179348 5941 factory.go:656] Stopping watch factory\\\\nI1215 12:17:54.179363 5941 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1215 12:17:54.179372 5941 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1215 12:17:54.179378 5941 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1215 12:17:54.179428 5941 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179767 5941 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1215 12:17:54.180338 5941 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.830129 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.842018 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.852808 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.867479 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.879803 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.886554 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.886763 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.886890 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.886979 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.887195 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:57Z","lastTransitionTime":"2025-12-15T12:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.892472 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.907204 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.921678 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.936445 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.940673 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-9kh99"] Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.941170 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:17:57 crc kubenswrapper[4719]: E1215 12:17:57.941253 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.947041 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.956469 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.967772 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.984570 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:54Z\\\",\\\"message\\\":\\\"mers/factory.go:160\\\\nI1215 12:17:54.179259 5941 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179277 5941 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1215 12:17:54.179294 5941 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:54.179302 5941 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1215 12:17:54.179328 5941 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1215 12:17:54.179332 5941 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1215 12:17:54.179348 5941 factory.go:656] Stopping watch factory\\\\nI1215 12:17:54.179363 5941 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1215 12:17:54.179372 5941 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1215 12:17:54.179378 5941 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1215 12:17:54.179428 5941 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179767 5941 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1215 12:17:54.180338 5941 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\":17:56.741242 6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1215 12:17:56.741250 6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1215 12:17:56.741257 6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI1215 12:17:56.741264 6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI1215 12:17:56.741273 6093 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1215 12:17:56.741101 6093 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741416 6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:56.741467 6093 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741504 6093 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741688 6093 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741162 6093 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.989428 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.989595 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.989678 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.989760 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.989844 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:57Z","lastTransitionTime":"2025-12-15T12:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:57 crc kubenswrapper[4719]: I1215 12:17:57.998205 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:57Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.008249 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvcvn\" (UniqueName: \"kubernetes.io/projected/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-kube-api-access-fvcvn\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.008635 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.010523 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.026185 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.038524 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.052107 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.061724 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.078386 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.089827 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.093243 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.093283 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.093294 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.093311 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.093322 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:58Z","lastTransitionTime":"2025-12-15T12:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.101103 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.109120 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.109260 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvcvn\" (UniqueName: \"kubernetes.io/projected/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-kube-api-access-fvcvn\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:17:58 crc kubenswrapper[4719]: E1215 12:17:58.109317 4719 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:17:58 crc kubenswrapper[4719]: E1215 12:17:58.109463 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs podName:0f0b6df0-0612-46c9-83e3-b96f0ec5978d nodeName:}" failed. No retries permitted until 2025-12-15 12:17:58.609448919 +0000 UTC m=+39.551741949 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs") pod "network-metrics-daemon-9kh99" (UID: "0f0b6df0-0612-46c9-83e3-b96f0ec5978d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.112040 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.131014 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvcvn\" (UniqueName: \"kubernetes.io/projected/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-kube-api-access-fvcvn\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.140831 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.155071 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.176126 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.195474 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.195513 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.195522 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.195535 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.195543 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:58Z","lastTransitionTime":"2025-12-15T12:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.197080 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.211964 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:58Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.297925 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.298270 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.298334 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.298394 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.298490 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:58Z","lastTransitionTime":"2025-12-15T12:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.401378 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.401431 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.401447 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.401462 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.401471 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:58Z","lastTransitionTime":"2025-12-15T12:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.475660 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.475699 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:17:58 crc kubenswrapper[4719]: E1215 12:17:58.475824 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.475672 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:17:58 crc kubenswrapper[4719]: E1215 12:17:58.475993 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:17:58 crc kubenswrapper[4719]: E1215 12:17:58.476118 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.503888 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.504176 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.504302 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.504392 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.504508 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:58Z","lastTransitionTime":"2025-12-15T12:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.608082 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.608452 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.608608 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.608816 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.608979 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:58Z","lastTransitionTime":"2025-12-15T12:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.612677 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:17:58 crc kubenswrapper[4719]: E1215 12:17:58.612925 4719 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:17:58 crc kubenswrapper[4719]: E1215 12:17:58.613023 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs podName:0f0b6df0-0612-46c9-83e3-b96f0ec5978d nodeName:}" failed. No retries permitted until 2025-12-15 12:17:59.61299445 +0000 UTC m=+40.555287530 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs") pod "network-metrics-daemon-9kh99" (UID: "0f0b6df0-0612-46c9-83e3-b96f0ec5978d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.711935 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.713226 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.713320 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.713422 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.713512 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:58Z","lastTransitionTime":"2025-12-15T12:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.774478 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/1.log" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.816551 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.816591 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.816601 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.816626 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.816638 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:58Z","lastTransitionTime":"2025-12-15T12:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.919269 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.919303 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.919312 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.919326 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:58 crc kubenswrapper[4719]: I1215 12:17:58.919336 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:58Z","lastTransitionTime":"2025-12-15T12:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.021541 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.021592 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.021602 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.021616 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.021625 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:59Z","lastTransitionTime":"2025-12-15T12:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.124087 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.124165 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.124178 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.124197 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.124210 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:59Z","lastTransitionTime":"2025-12-15T12:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.227365 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.227405 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.227418 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.227434 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.227445 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:59Z","lastTransitionTime":"2025-12-15T12:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.330070 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.330127 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.330141 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.330162 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.330173 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:59Z","lastTransitionTime":"2025-12-15T12:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.432397 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.432763 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.432843 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.432938 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.433006 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:59Z","lastTransitionTime":"2025-12-15T12:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.476006 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:17:59 crc kubenswrapper[4719]: E1215 12:17:59.476478 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.497682 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.516806 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.529520 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.535299 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.535329 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.535342 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.535359 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.535373 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:59Z","lastTransitionTime":"2025-12-15T12:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.554436 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.576588 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.589407 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.605757 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.618412 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.625697 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:17:59 crc kubenswrapper[4719]: E1215 12:17:59.625825 4719 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:17:59 crc kubenswrapper[4719]: E1215 12:17:59.626059 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs podName:0f0b6df0-0612-46c9-83e3-b96f0ec5978d nodeName:}" failed. No retries permitted until 2025-12-15 12:18:01.626044837 +0000 UTC m=+42.568337867 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs") pod "network-metrics-daemon-9kh99" (UID: "0f0b6df0-0612-46c9-83e3-b96f0ec5978d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.630538 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.638445 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.638478 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.638490 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.638505 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.638517 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:59Z","lastTransitionTime":"2025-12-15T12:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.642553 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.656274 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.671647 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.680762 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.693000 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.708117 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.730920 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:54Z\\\",\\\"message\\\":\\\"mers/factory.go:160\\\\nI1215 12:17:54.179259 5941 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179277 5941 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1215 12:17:54.179294 5941 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:54.179302 5941 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1215 12:17:54.179328 5941 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1215 12:17:54.179332 5941 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1215 12:17:54.179348 5941 factory.go:656] Stopping watch factory\\\\nI1215 12:17:54.179363 5941 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1215 12:17:54.179372 5941 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1215 12:17:54.179378 5941 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1215 12:17:54.179428 5941 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179767 5941 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1215 12:17:54.180338 5941 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\":17:56.741242 6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1215 12:17:56.741250 6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1215 12:17:56.741257 6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI1215 12:17:56.741264 6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI1215 12:17:56.741273 6093 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1215 12:17:56.741101 6093 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741416 6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:56.741467 6093 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741504 6093 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741688 6093 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741162 6093 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:17:59Z is after 2025-08-24T17:21:41Z" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.740685 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.740726 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.740738 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.740759 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.740771 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:59Z","lastTransitionTime":"2025-12-15T12:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.844208 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.845032 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.845221 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.845584 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.845745 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:59Z","lastTransitionTime":"2025-12-15T12:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.948473 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.948531 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.948543 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.948560 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:17:59 crc kubenswrapper[4719]: I1215 12:17:59.948572 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:17:59Z","lastTransitionTime":"2025-12-15T12:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.051889 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.051960 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.051980 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.052005 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.052025 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:00Z","lastTransitionTime":"2025-12-15T12:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.155378 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.155418 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.155432 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.155447 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.155457 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:00Z","lastTransitionTime":"2025-12-15T12:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.258155 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.258201 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.258210 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.258225 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.258234 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:00Z","lastTransitionTime":"2025-12-15T12:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.361097 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.361130 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.361142 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.361158 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.361169 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:00Z","lastTransitionTime":"2025-12-15T12:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.463693 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.463749 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.463763 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.463783 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.463794 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:00Z","lastTransitionTime":"2025-12-15T12:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.475832 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.475904 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:00 crc kubenswrapper[4719]: E1215 12:18:00.475949 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:00 crc kubenswrapper[4719]: E1215 12:18:00.475997 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.475832 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:00 crc kubenswrapper[4719]: E1215 12:18:00.476198 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.566231 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.566278 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.566292 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.566307 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.566333 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:00Z","lastTransitionTime":"2025-12-15T12:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.668378 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.668458 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.668475 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.668496 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.668512 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:00Z","lastTransitionTime":"2025-12-15T12:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.770843 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.770899 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.770910 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.770925 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.770934 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:00Z","lastTransitionTime":"2025-12-15T12:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.874321 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.874665 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.874734 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.874823 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.874919 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:00Z","lastTransitionTime":"2025-12-15T12:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.977995 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.978041 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.978050 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.978063 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:00 crc kubenswrapper[4719]: I1215 12:18:00.978073 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:00Z","lastTransitionTime":"2025-12-15T12:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.080834 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.080896 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.080908 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.080924 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.080936 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:01Z","lastTransitionTime":"2025-12-15T12:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.182999 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.183061 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.183074 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.183104 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.183116 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:01Z","lastTransitionTime":"2025-12-15T12:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.285797 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.286212 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.286313 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.286463 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.286601 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:01Z","lastTransitionTime":"2025-12-15T12:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.390252 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.390322 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.390332 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.390353 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.390365 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:01Z","lastTransitionTime":"2025-12-15T12:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.475575 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:01 crc kubenswrapper[4719]: E1215 12:18:01.475837 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.492511 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.492569 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.492582 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.492600 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.492616 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:01Z","lastTransitionTime":"2025-12-15T12:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.595810 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.596005 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.596037 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.596067 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.596089 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:01Z","lastTransitionTime":"2025-12-15T12:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.642645 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:01 crc kubenswrapper[4719]: E1215 12:18:01.642996 4719 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:18:01 crc kubenswrapper[4719]: E1215 12:18:01.643162 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs podName:0f0b6df0-0612-46c9-83e3-b96f0ec5978d nodeName:}" failed. No retries permitted until 2025-12-15 12:18:05.643124886 +0000 UTC m=+46.585417976 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs") pod "network-metrics-daemon-9kh99" (UID: "0f0b6df0-0612-46c9-83e3-b96f0ec5978d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.699049 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.699112 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.699140 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.699171 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.699190 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:01Z","lastTransitionTime":"2025-12-15T12:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.801706 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.802233 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.802242 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.802257 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.802266 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:01Z","lastTransitionTime":"2025-12-15T12:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.905358 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.905430 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.905456 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.905487 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:01 crc kubenswrapper[4719]: I1215 12:18:01.905510 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:01Z","lastTransitionTime":"2025-12-15T12:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.009287 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.009349 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.009362 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.009381 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.009395 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:02Z","lastTransitionTime":"2025-12-15T12:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.113637 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.113703 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.113726 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.113754 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.113774 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:02Z","lastTransitionTime":"2025-12-15T12:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.216203 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.216255 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.216267 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.216285 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.216299 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:02Z","lastTransitionTime":"2025-12-15T12:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.319091 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.319144 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.319156 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.319176 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.319188 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:02Z","lastTransitionTime":"2025-12-15T12:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.422147 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.422208 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.422231 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.422259 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.422282 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:02Z","lastTransitionTime":"2025-12-15T12:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.475460 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.475629 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:02 crc kubenswrapper[4719]: E1215 12:18:02.475733 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.475927 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:02 crc kubenswrapper[4719]: E1215 12:18:02.476208 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:02 crc kubenswrapper[4719]: E1215 12:18:02.476290 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.525556 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.525596 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.525608 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.525623 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.525634 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:02Z","lastTransitionTime":"2025-12-15T12:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.628792 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.628839 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.628875 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.628900 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.628915 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:02Z","lastTransitionTime":"2025-12-15T12:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.732446 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.732548 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.732575 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.732607 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.732632 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:02Z","lastTransitionTime":"2025-12-15T12:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.835669 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.835738 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.835762 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.835793 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.835816 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:02Z","lastTransitionTime":"2025-12-15T12:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.939028 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.939084 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.939101 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.939123 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:02 crc kubenswrapper[4719]: I1215 12:18:02.939141 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:02Z","lastTransitionTime":"2025-12-15T12:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.001808 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.001921 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.001949 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.001980 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.002004 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: E1215 12:18:03.030382 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:03Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.036491 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.036554 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.036577 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.036753 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.036828 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: E1215 12:18:03.059741 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:03Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.063695 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.063753 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.063772 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.063790 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.063804 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: E1215 12:18:03.077365 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:03Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.082841 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.082949 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.082961 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.082978 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.082990 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: E1215 12:18:03.097766 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:03Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.103334 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.103386 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.103399 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.103417 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.103429 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: E1215 12:18:03.117670 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:03Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:03 crc kubenswrapper[4719]: E1215 12:18:03.117828 4719 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.119719 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.119761 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.119775 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.119797 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.119813 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.224220 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.224366 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.224391 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.224420 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.224441 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.327938 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.327995 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.328011 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.328034 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.328052 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.431309 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.431499 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.431528 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.431561 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.431596 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.475098 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:03 crc kubenswrapper[4719]: E1215 12:18:03.475280 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.534327 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.534649 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.534744 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.534835 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.534952 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.638034 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.638538 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.638748 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.638998 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.639215 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.742210 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.742269 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.742281 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.742296 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.742305 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.845585 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.845640 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.845655 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.845678 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.845695 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.948569 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.948637 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.948654 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.948689 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:03 crc kubenswrapper[4719]: I1215 12:18:03.948702 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:03Z","lastTransitionTime":"2025-12-15T12:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.051164 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.051211 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.051220 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.051234 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.051243 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:04Z","lastTransitionTime":"2025-12-15T12:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.153775 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.153837 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.153874 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.153892 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.153903 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:04Z","lastTransitionTime":"2025-12-15T12:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.257047 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.257108 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.257124 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.257147 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.257166 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:04Z","lastTransitionTime":"2025-12-15T12:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.360135 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.360200 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.360218 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.360242 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.360271 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:04Z","lastTransitionTime":"2025-12-15T12:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.463703 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.463747 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.463756 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.463772 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.463782 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:04Z","lastTransitionTime":"2025-12-15T12:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.475624 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.475643 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.475744 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:04 crc kubenswrapper[4719]: E1215 12:18:04.475897 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:04 crc kubenswrapper[4719]: E1215 12:18:04.476027 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:04 crc kubenswrapper[4719]: E1215 12:18:04.476146 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.566798 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.566848 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.566887 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.566907 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.566919 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:04Z","lastTransitionTime":"2025-12-15T12:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.669107 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.669188 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.669212 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.669242 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.669263 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:04Z","lastTransitionTime":"2025-12-15T12:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.771796 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.771844 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.771875 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.771891 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.771901 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:04Z","lastTransitionTime":"2025-12-15T12:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.875140 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.875206 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.875227 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.875252 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.875270 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:04Z","lastTransitionTime":"2025-12-15T12:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.979031 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.979097 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.979122 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.979151 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:04 crc kubenswrapper[4719]: I1215 12:18:04.979175 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:04Z","lastTransitionTime":"2025-12-15T12:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.083177 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.083245 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.083263 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.083286 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.083306 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:05Z","lastTransitionTime":"2025-12-15T12:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.186059 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.186126 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.186148 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.186179 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.186206 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:05Z","lastTransitionTime":"2025-12-15T12:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.289618 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.289685 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.289707 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.289735 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.289756 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:05Z","lastTransitionTime":"2025-12-15T12:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.391816 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.391877 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.391887 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.391903 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.391915 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:05Z","lastTransitionTime":"2025-12-15T12:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.476178 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:05 crc kubenswrapper[4719]: E1215 12:18:05.476430 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.494658 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.494699 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.494707 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.494720 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.494730 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:05Z","lastTransitionTime":"2025-12-15T12:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.598015 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.598077 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.598097 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.598121 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.598142 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:05Z","lastTransitionTime":"2025-12-15T12:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.701258 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.701299 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.701310 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.701327 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.701339 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:05Z","lastTransitionTime":"2025-12-15T12:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.714521 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:05 crc kubenswrapper[4719]: E1215 12:18:05.714720 4719 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:18:05 crc kubenswrapper[4719]: E1215 12:18:05.714802 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs podName:0f0b6df0-0612-46c9-83e3-b96f0ec5978d nodeName:}" failed. No retries permitted until 2025-12-15 12:18:13.714784121 +0000 UTC m=+54.657077161 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs") pod "network-metrics-daemon-9kh99" (UID: "0f0b6df0-0612-46c9-83e3-b96f0ec5978d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.803510 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.803561 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.803576 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.803596 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.803609 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:05Z","lastTransitionTime":"2025-12-15T12:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.907920 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.907991 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.908009 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.908034 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:05 crc kubenswrapper[4719]: I1215 12:18:05.908052 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:05Z","lastTransitionTime":"2025-12-15T12:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.011178 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.011242 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.011256 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.011276 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.011290 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:06Z","lastTransitionTime":"2025-12-15T12:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.114212 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.114274 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.114293 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.114308 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.114317 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:06Z","lastTransitionTime":"2025-12-15T12:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.216703 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.216807 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.216825 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.216843 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.216873 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:06Z","lastTransitionTime":"2025-12-15T12:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.319806 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.319926 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.319955 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.319988 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.320013 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:06Z","lastTransitionTime":"2025-12-15T12:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.423349 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.423406 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.423430 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.423459 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.423478 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:06Z","lastTransitionTime":"2025-12-15T12:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.475093 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.475093 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:06 crc kubenswrapper[4719]: E1215 12:18:06.475296 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:06 crc kubenswrapper[4719]: E1215 12:18:06.475394 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.475104 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:06 crc kubenswrapper[4719]: E1215 12:18:06.475558 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.525802 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.525839 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.525851 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.525889 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.525904 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:06Z","lastTransitionTime":"2025-12-15T12:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.628262 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.628355 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.628364 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.628383 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.628392 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:06Z","lastTransitionTime":"2025-12-15T12:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.731637 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.731724 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.731737 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.731752 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.731794 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:06Z","lastTransitionTime":"2025-12-15T12:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.834391 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.834456 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.834472 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.834497 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.834515 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:06Z","lastTransitionTime":"2025-12-15T12:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.936692 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.936760 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.936801 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.936833 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:06 crc kubenswrapper[4719]: I1215 12:18:06.936903 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:06Z","lastTransitionTime":"2025-12-15T12:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.040582 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.040670 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.040686 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.040709 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.040726 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:07Z","lastTransitionTime":"2025-12-15T12:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.143768 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.143845 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.143886 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.143909 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.143926 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:07Z","lastTransitionTime":"2025-12-15T12:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.248101 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.248180 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.248194 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.248213 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.248230 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:07Z","lastTransitionTime":"2025-12-15T12:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.350273 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.350316 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.350325 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.350337 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.350345 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:07Z","lastTransitionTime":"2025-12-15T12:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.453484 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.453546 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.453566 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.453589 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.453606 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:07Z","lastTransitionTime":"2025-12-15T12:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.475799 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:07 crc kubenswrapper[4719]: E1215 12:18:07.476022 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.556913 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.556954 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.556967 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.556985 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.556997 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:07Z","lastTransitionTime":"2025-12-15T12:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.659891 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.659945 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.659981 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.660012 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.660029 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:07Z","lastTransitionTime":"2025-12-15T12:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.763279 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.763336 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.763347 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.763365 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.763376 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:07Z","lastTransitionTime":"2025-12-15T12:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.866259 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.866302 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.866311 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.866328 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.866338 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:07Z","lastTransitionTime":"2025-12-15T12:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.968727 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.968801 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.968828 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.968903 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:07 crc kubenswrapper[4719]: I1215 12:18:07.968932 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:07Z","lastTransitionTime":"2025-12-15T12:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.072042 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.072095 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.072110 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.072130 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.072147 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:08Z","lastTransitionTime":"2025-12-15T12:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.177232 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.178178 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.178200 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.178223 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.178239 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:08Z","lastTransitionTime":"2025-12-15T12:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.281552 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.281607 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.281622 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.281645 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.281661 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:08Z","lastTransitionTime":"2025-12-15T12:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.384430 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.384472 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.384483 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.384503 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.384512 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:08Z","lastTransitionTime":"2025-12-15T12:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.476063 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.476183 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:08 crc kubenswrapper[4719]: E1215 12:18:08.476252 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.476344 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:08 crc kubenswrapper[4719]: E1215 12:18:08.476538 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:08 crc kubenswrapper[4719]: E1215 12:18:08.476656 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.487186 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.487237 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.487255 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.487278 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.487296 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:08Z","lastTransitionTime":"2025-12-15T12:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.590112 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.590648 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.590908 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.591162 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.591374 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:08Z","lastTransitionTime":"2025-12-15T12:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.694722 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.694788 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.694811 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.694842 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.694896 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:08Z","lastTransitionTime":"2025-12-15T12:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.798539 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.798603 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.798623 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.798655 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.798677 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:08Z","lastTransitionTime":"2025-12-15T12:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.902360 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.902407 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.902418 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.902434 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:08 crc kubenswrapper[4719]: I1215 12:18:08.902446 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:08Z","lastTransitionTime":"2025-12-15T12:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.005605 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.005644 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.005656 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.005673 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.005685 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:09Z","lastTransitionTime":"2025-12-15T12:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.108878 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.108919 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.108929 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.108946 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.108957 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:09Z","lastTransitionTime":"2025-12-15T12:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.156441 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.165519 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.173214 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.186885 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.201424 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.211678 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.211718 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.211731 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.211750 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.211763 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:09Z","lastTransitionTime":"2025-12-15T12:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.218005 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.236454 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.250560 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.262941 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.277210 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.295945 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.314417 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.314468 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.314483 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.314503 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.314518 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:09Z","lastTransitionTime":"2025-12-15T12:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.318601 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:54Z\\\",\\\"message\\\":\\\"mers/factory.go:160\\\\nI1215 12:17:54.179259 5941 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179277 5941 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1215 12:17:54.179294 5941 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:54.179302 5941 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1215 12:17:54.179328 5941 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1215 12:17:54.179332 5941 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1215 12:17:54.179348 5941 factory.go:656] Stopping watch factory\\\\nI1215 12:17:54.179363 5941 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1215 12:17:54.179372 5941 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1215 12:17:54.179378 5941 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1215 12:17:54.179428 5941 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179767 5941 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1215 12:17:54.180338 5941 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\":17:56.741242 6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1215 12:17:56.741250 6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1215 12:17:56.741257 6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI1215 12:17:56.741264 6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI1215 12:17:56.741273 6093 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1215 12:17:56.741101 6093 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741416 6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:56.741467 6093 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741504 6093 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741688 6093 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741162 6093 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.334968 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.350333 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.364459 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.383933 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.402494 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.416823 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.416882 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.416893 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.416910 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.416921 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:09Z","lastTransitionTime":"2025-12-15T12:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.420052 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.475834 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:09 crc kubenswrapper[4719]: E1215 12:18:09.476044 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.495173 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.508395 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.520486 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.520543 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.520558 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.520579 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.520594 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:09Z","lastTransitionTime":"2025-12-15T12:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.530757 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.545045 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.555917 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.575272 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.590171 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.598532 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.609150 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.618316 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.622411 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.622439 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.622463 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.622480 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.622490 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:09Z","lastTransitionTime":"2025-12-15T12:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.630378 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.640030 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.650796 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.661232 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.676116 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.692982 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f596a50ece3a3f6b0a99a6dac648606726b8526a387160fdeb33f77d988e4cc8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:54Z\\\",\\\"message\\\":\\\"mers/factory.go:160\\\\nI1215 12:17:54.179259 5941 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179277 5941 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1215 12:17:54.179294 5941 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:54.179302 5941 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1215 12:17:54.179328 5941 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1215 12:17:54.179332 5941 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1215 12:17:54.179348 5941 factory.go:656] Stopping watch factory\\\\nI1215 12:17:54.179363 5941 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1215 12:17:54.179372 5941 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1215 12:17:54.179378 5941 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1215 12:17:54.179428 5941 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:54.179767 5941 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1215 12:17:54.180338 5941 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\":17:56.741242 6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1215 12:17:56.741250 6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1215 12:17:56.741257 6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI1215 12:17:56.741264 6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI1215 12:17:56.741273 6093 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1215 12:17:56.741101 6093 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741416 6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:56.741467 6093 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741504 6093 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741688 6093 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741162 6093 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.707650 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:09Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.724303 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.724348 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.724359 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.724374 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.724382 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:09Z","lastTransitionTime":"2025-12-15T12:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.826493 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.826531 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.826540 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.826556 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.826567 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:09Z","lastTransitionTime":"2025-12-15T12:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.929416 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.929491 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.929534 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.929552 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:09 crc kubenswrapper[4719]: I1215 12:18:09.929564 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:09Z","lastTransitionTime":"2025-12-15T12:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.032371 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.032409 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.032419 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.032430 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.032439 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:10Z","lastTransitionTime":"2025-12-15T12:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.135076 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.135168 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.135200 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.135231 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.135256 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:10Z","lastTransitionTime":"2025-12-15T12:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.239058 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.239107 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.239119 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.239136 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.239148 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:10Z","lastTransitionTime":"2025-12-15T12:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.342781 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.342832 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.342843 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.342881 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.342896 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:10Z","lastTransitionTime":"2025-12-15T12:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.445554 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.445601 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.445614 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.445632 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.445644 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:10Z","lastTransitionTime":"2025-12-15T12:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.475534 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.475566 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.475596 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.475693 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.475834 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.475994 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.547930 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.548012 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.548028 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.548047 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.548062 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:10Z","lastTransitionTime":"2025-12-15T12:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.563750 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.563957 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564010 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:18:42.563975736 +0000 UTC m=+83.506268776 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.564077 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564098 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564118 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564133 4719 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.564146 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564192 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-15 12:18:42.564172602 +0000 UTC m=+83.506465652 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.564217 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564326 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564343 4719 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564359 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564375 4719 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564399 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:18:42.564384549 +0000 UTC m=+83.506677649 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564425 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:18:42.56440872 +0000 UTC m=+83.506701800 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564503 4719 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:18:10 crc kubenswrapper[4719]: E1215 12:18:10.564601 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:18:42.564582505 +0000 UTC m=+83.506875525 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.650947 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.651006 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.651021 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.651043 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.651058 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:10Z","lastTransitionTime":"2025-12-15T12:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.753235 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.753280 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.753292 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.753309 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.753320 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:10Z","lastTransitionTime":"2025-12-15T12:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.855755 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.855799 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.855813 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.855832 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.855885 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:10Z","lastTransitionTime":"2025-12-15T12:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.958454 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.958529 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.958550 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.958582 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:10 crc kubenswrapper[4719]: I1215 12:18:10.958604 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:10Z","lastTransitionTime":"2025-12-15T12:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.061971 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.062015 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.062026 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.062043 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.062056 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:11Z","lastTransitionTime":"2025-12-15T12:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.168488 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.168548 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.168564 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.168584 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.168602 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:11Z","lastTransitionTime":"2025-12-15T12:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.270995 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.271032 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.271042 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.271060 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.271071 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:11Z","lastTransitionTime":"2025-12-15T12:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.373691 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.373731 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.373742 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.373757 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.373769 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:11Z","lastTransitionTime":"2025-12-15T12:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.475075 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:11 crc kubenswrapper[4719]: E1215 12:18:11.475254 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.476350 4719 scope.go:117] "RemoveContainer" containerID="a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.476559 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.476613 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.476632 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.476657 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.476675 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:11Z","lastTransitionTime":"2025-12-15T12:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.492741 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.506174 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.520831 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.536937 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.551747 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.576347 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.581013 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.581058 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.581067 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.581080 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.581089 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:11Z","lastTransitionTime":"2025-12-15T12:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.592358 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.605329 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.618624 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.633224 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.660184 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\":17:56.741242 6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1215 12:17:56.741250 6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1215 12:17:56.741257 6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI1215 12:17:56.741264 6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI1215 12:17:56.741273 6093 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1215 12:17:56.741101 6093 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741416 6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:56.741467 6093 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741504 6093 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741688 6093 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741162 6093 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.674804 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.683456 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.683494 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.683505 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.683520 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.683532 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:11Z","lastTransitionTime":"2025-12-15T12:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.686383 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.699216 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.710654 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.721902 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.734892 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.786218 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.786247 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.786255 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.786269 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.786278 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:11Z","lastTransitionTime":"2025-12-15T12:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.828537 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/1.log" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.831662 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.831819 4719 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.845065 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.856624 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.868734 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.880739 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.891578 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.891618 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.891629 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.891647 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.891660 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:11Z","lastTransitionTime":"2025-12-15T12:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.894395 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.915247 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.935831 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.949986 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.961956 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.979559 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.993793 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:11Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.993955 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.993997 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.994012 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.994032 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:11 crc kubenswrapper[4719]: I1215 12:18:11.994046 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:11Z","lastTransitionTime":"2025-12-15T12:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.009690 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.022536 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.037534 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.037549 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.061032 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\":17:56.741242 6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1215 12:17:56.741250 6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1215 12:17:56.741257 6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI1215 12:17:56.741264 6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI1215 12:17:56.741273 6093 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1215 12:17:56.741101 6093 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741416 6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:56.741467 6093 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741504 6093 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741688 6093 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741162 6093 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.081556 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.096116 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.096157 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.096168 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.096182 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.096192 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:12Z","lastTransitionTime":"2025-12-15T12:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.096703 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.198665 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.198695 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.198703 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.198717 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.198726 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:12Z","lastTransitionTime":"2025-12-15T12:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.300711 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.300758 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.300771 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.300790 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.300800 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:12Z","lastTransitionTime":"2025-12-15T12:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.403789 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.403841 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.403892 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.403925 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.403949 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:12Z","lastTransitionTime":"2025-12-15T12:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.476024 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.476044 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.476048 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:12 crc kubenswrapper[4719]: E1215 12:18:12.476171 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:12 crc kubenswrapper[4719]: E1215 12:18:12.476278 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:12 crc kubenswrapper[4719]: E1215 12:18:12.476374 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.506781 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.506832 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.506842 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.506875 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.506887 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:12Z","lastTransitionTime":"2025-12-15T12:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.609674 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.609725 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.609737 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.609757 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.609770 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:12Z","lastTransitionTime":"2025-12-15T12:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.711699 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.711739 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.711752 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.711769 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.711780 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:12Z","lastTransitionTime":"2025-12-15T12:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.815501 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.815549 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.815566 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.815603 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.815621 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:12Z","lastTransitionTime":"2025-12-15T12:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.838511 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/2.log" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.839470 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/1.log" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.842459 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3" exitCode=1 Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.842500 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.842530 4719 scope.go:117] "RemoveContainer" containerID="a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.843663 4719 scope.go:117] "RemoveContainer" containerID="a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3" Dec 15 12:18:12 crc kubenswrapper[4719]: E1215 12:18:12.844136 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.898111 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.919746 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.919789 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.919799 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.919815 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.919826 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:12Z","lastTransitionTime":"2025-12-15T12:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.927809 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.941180 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.956169 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.968060 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.979445 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:12 crc kubenswrapper[4719]: I1215 12:18:12.990137 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.000496 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:12Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.011457 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.022014 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.022036 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.022044 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.022058 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.022066 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.022395 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.035150 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.045259 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.057085 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.070007 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.084652 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.103866 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6103172d12d9ef6fd8f0522b95923ec6cac2ca327d6c900eeb8f86a92a1a6e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\":17:56.741242 6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1215 12:17:56.741250 6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1215 12:17:56.741257 6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI1215 12:17:56.741264 6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI1215 12:17:56.741273 6093 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1215 12:17:56.741101 6093 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741416 6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1215 12:17:56.741467 6093 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741504 6093 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1215 12:17:56.741688 6093 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1215 12:17:56.741162 6093 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:12Z\\\",\\\"message\\\":\\\"3047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00720970f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.117204 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.124331 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.124371 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.124382 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.124400 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.124413 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.207274 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.207512 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.207523 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.207540 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.207552 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: E1215 12:18:13.221264 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.225037 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.225105 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.225120 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.225145 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.225162 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: E1215 12:18:13.237648 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.241540 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.241657 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.241718 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.241783 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.241841 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: E1215 12:18:13.255405 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.258898 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.258927 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.258937 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.258954 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.258966 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: E1215 12:18:13.270465 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.273756 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.273810 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.273821 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.273837 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.273848 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: E1215 12:18:13.285349 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: E1215 12:18:13.285889 4719 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.287639 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.287666 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.287676 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.287691 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.287701 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.390317 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.390589 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.390741 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.390811 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.390915 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.475446 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:13 crc kubenswrapper[4719]: E1215 12:18:13.475589 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.493754 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.494180 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.494306 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.494409 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.494486 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.598953 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.599000 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.599012 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.599031 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.599043 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.706910 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.706951 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.706964 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.706984 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.706998 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.798838 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:13 crc kubenswrapper[4719]: E1215 12:18:13.799096 4719 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:18:13 crc kubenswrapper[4719]: E1215 12:18:13.799192 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs podName:0f0b6df0-0612-46c9-83e3-b96f0ec5978d nodeName:}" failed. No retries permitted until 2025-12-15 12:18:29.799163386 +0000 UTC m=+70.741456416 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs") pod "network-metrics-daemon-9kh99" (UID: "0f0b6df0-0612-46c9-83e3-b96f0ec5978d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.819983 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.820035 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.820046 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.820063 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.820074 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.848625 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/2.log" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.852930 4719 scope.go:117] "RemoveContainer" containerID="a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3" Dec 15 12:18:13 crc kubenswrapper[4719]: E1215 12:18:13.853100 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.869523 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.884643 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.902631 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.922723 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.922803 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.922829 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.922842 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.922887 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.922903 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:13Z","lastTransitionTime":"2025-12-15T12:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.936318 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.956411 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.973574 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:13 crc kubenswrapper[4719]: I1215 12:18:13.990280 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:13Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.013613 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:12Z\\\",\\\"message\\\":\\\"3047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00720970f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:14Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.025612 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.025650 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.025662 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.025679 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.025690 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:14Z","lastTransitionTime":"2025-12-15T12:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.029344 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:14Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.042148 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:14Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.057311 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:14Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.071784 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:14Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.084124 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:14Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.096733 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:14Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.108765 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:14Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.126421 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:14Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.127605 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.127650 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.127663 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.127681 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.127693 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:14Z","lastTransitionTime":"2025-12-15T12:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.231682 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.232173 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.232283 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.232378 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.232462 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:14Z","lastTransitionTime":"2025-12-15T12:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.335191 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.335229 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.335241 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.335259 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.335271 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:14Z","lastTransitionTime":"2025-12-15T12:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.438498 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.438582 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.438604 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.438634 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.438654 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:14Z","lastTransitionTime":"2025-12-15T12:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.475526 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.475589 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.475588 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:14 crc kubenswrapper[4719]: E1215 12:18:14.476073 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:14 crc kubenswrapper[4719]: E1215 12:18:14.476157 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:14 crc kubenswrapper[4719]: E1215 12:18:14.475837 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.541327 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.541365 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.541376 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.541390 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.541402 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:14Z","lastTransitionTime":"2025-12-15T12:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.644340 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.644612 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.644734 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.644827 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.644936 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:14Z","lastTransitionTime":"2025-12-15T12:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.747636 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.747674 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.747682 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.747697 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.747707 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:14Z","lastTransitionTime":"2025-12-15T12:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.850663 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.850706 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.850717 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.850733 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.850744 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:14Z","lastTransitionTime":"2025-12-15T12:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.952913 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.952953 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.952964 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.952979 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:14 crc kubenswrapper[4719]: I1215 12:18:14.952990 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:14Z","lastTransitionTime":"2025-12-15T12:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.054994 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.055027 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.055038 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.055059 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.055073 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:15Z","lastTransitionTime":"2025-12-15T12:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.157178 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.157718 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.157845 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.158060 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.158221 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:15Z","lastTransitionTime":"2025-12-15T12:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.260667 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.260702 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.260710 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.260724 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.260732 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:15Z","lastTransitionTime":"2025-12-15T12:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.362845 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.363145 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.363227 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.363305 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.363376 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:15Z","lastTransitionTime":"2025-12-15T12:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.465959 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.466002 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.466014 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.466029 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.466040 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:15Z","lastTransitionTime":"2025-12-15T12:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.475249 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:15 crc kubenswrapper[4719]: E1215 12:18:15.475362 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.568687 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.568721 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.568730 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.568744 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.568756 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:15Z","lastTransitionTime":"2025-12-15T12:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.670838 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.670896 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.670906 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.670920 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.670928 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:15Z","lastTransitionTime":"2025-12-15T12:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.775250 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.775288 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.775299 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.775315 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.775329 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:15Z","lastTransitionTime":"2025-12-15T12:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.877796 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.877870 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.877881 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.877895 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.877903 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:15Z","lastTransitionTime":"2025-12-15T12:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.981379 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.981533 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.981560 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.981589 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:15 crc kubenswrapper[4719]: I1215 12:18:15.981611 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:15Z","lastTransitionTime":"2025-12-15T12:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.084660 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.084718 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.084742 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.084770 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.084793 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:16Z","lastTransitionTime":"2025-12-15T12:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.187169 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.187225 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.187243 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.187267 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.187285 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:16Z","lastTransitionTime":"2025-12-15T12:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.289812 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.289850 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.289872 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.289886 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.289895 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:16Z","lastTransitionTime":"2025-12-15T12:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.392832 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.392876 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.392885 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.392899 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.392908 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:16Z","lastTransitionTime":"2025-12-15T12:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.475444 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.475562 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.475501 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:16 crc kubenswrapper[4719]: E1215 12:18:16.475697 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:16 crc kubenswrapper[4719]: E1215 12:18:16.475903 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:16 crc kubenswrapper[4719]: E1215 12:18:16.476054 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.495360 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.495421 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.495438 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.495464 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.495481 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:16Z","lastTransitionTime":"2025-12-15T12:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.598559 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.598612 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.598629 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.598649 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.598663 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:16Z","lastTransitionTime":"2025-12-15T12:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.702118 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.702170 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.702187 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.702211 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.702227 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:16Z","lastTransitionTime":"2025-12-15T12:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.805380 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.805470 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.805497 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.805530 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.805557 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:16Z","lastTransitionTime":"2025-12-15T12:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.908907 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.908939 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.908952 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.908969 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:16 crc kubenswrapper[4719]: I1215 12:18:16.908980 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:16Z","lastTransitionTime":"2025-12-15T12:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.013017 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.013118 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.013144 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.013175 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.013199 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:17Z","lastTransitionTime":"2025-12-15T12:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.115836 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.115889 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.115903 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.115920 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.115933 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:17Z","lastTransitionTime":"2025-12-15T12:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.218235 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.218272 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.218282 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.218298 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.218311 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:17Z","lastTransitionTime":"2025-12-15T12:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.320396 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.320449 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.320466 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.320490 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.320506 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:17Z","lastTransitionTime":"2025-12-15T12:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.422771 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.422822 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.422837 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.422889 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.422915 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:17Z","lastTransitionTime":"2025-12-15T12:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.476141 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:17 crc kubenswrapper[4719]: E1215 12:18:17.476357 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.525082 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.525140 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.525151 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.525168 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.525178 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:17Z","lastTransitionTime":"2025-12-15T12:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.628003 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.628069 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.628088 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.628113 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.628132 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:17Z","lastTransitionTime":"2025-12-15T12:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.730261 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.730331 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.730354 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.730379 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.730397 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:17Z","lastTransitionTime":"2025-12-15T12:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.833176 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.833216 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.833224 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.833237 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.833246 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:17Z","lastTransitionTime":"2025-12-15T12:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.939156 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.939224 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.939247 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.939275 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:17 crc kubenswrapper[4719]: I1215 12:18:17.939305 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:17Z","lastTransitionTime":"2025-12-15T12:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.042342 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.042391 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.042401 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.042416 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.042424 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:18Z","lastTransitionTime":"2025-12-15T12:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.145079 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.145106 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.145116 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.145129 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.145138 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:18Z","lastTransitionTime":"2025-12-15T12:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.247655 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.247686 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.247736 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.247753 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.247762 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:18Z","lastTransitionTime":"2025-12-15T12:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.349790 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.349841 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.349874 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.349893 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.349906 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:18Z","lastTransitionTime":"2025-12-15T12:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.452898 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.452937 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.452946 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.452960 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.452969 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:18Z","lastTransitionTime":"2025-12-15T12:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.476044 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:18 crc kubenswrapper[4719]: E1215 12:18:18.476150 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.476040 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.476044 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:18 crc kubenswrapper[4719]: E1215 12:18:18.476218 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:18 crc kubenswrapper[4719]: E1215 12:18:18.476388 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.555224 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.555260 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.555269 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.555284 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.555293 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:18Z","lastTransitionTime":"2025-12-15T12:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.658327 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.658407 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.658440 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.658469 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.658488 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:18Z","lastTransitionTime":"2025-12-15T12:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.761369 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.761444 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.761467 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.761497 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.761518 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:18Z","lastTransitionTime":"2025-12-15T12:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.864247 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.864276 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.864284 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.864296 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.864305 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:18Z","lastTransitionTime":"2025-12-15T12:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.967401 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.967480 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.967503 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.967538 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:18 crc kubenswrapper[4719]: I1215 12:18:18.967559 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:18Z","lastTransitionTime":"2025-12-15T12:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.070684 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.070772 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.070792 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.070819 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.070848 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:19Z","lastTransitionTime":"2025-12-15T12:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.173609 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.173674 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.173697 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.173730 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.173751 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:19Z","lastTransitionTime":"2025-12-15T12:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.277269 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.277301 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.277311 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.277327 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.277336 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:19Z","lastTransitionTime":"2025-12-15T12:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.380312 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.380378 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.380403 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.380431 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.380452 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:19Z","lastTransitionTime":"2025-12-15T12:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.475429 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:19 crc kubenswrapper[4719]: E1215 12:18:19.475590 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.493229 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.493304 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.493331 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.493353 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.493393 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:19Z","lastTransitionTime":"2025-12-15T12:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.498221 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.516891 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.527645 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.539249 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.551296 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.569224 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.585886 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.596528 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.596781 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.596848 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.596973 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.597038 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:19Z","lastTransitionTime":"2025-12-15T12:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.600678 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.617954 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:12Z\\\",\\\"message\\\":\\\"3047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00720970f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.629816 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.641696 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.653122 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.663184 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.672765 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.684545 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.693987 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.699829 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.700056 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.700178 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.700267 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.700352 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:19Z","lastTransitionTime":"2025-12-15T12:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.706519 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:19Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.804267 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.804566 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.804656 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.804776 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.804894 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:19Z","lastTransitionTime":"2025-12-15T12:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.907803 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.908003 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.908033 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.908108 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:19 crc kubenswrapper[4719]: I1215 12:18:19.908135 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:19Z","lastTransitionTime":"2025-12-15T12:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.010831 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.010924 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.010941 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.010962 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.010979 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:20Z","lastTransitionTime":"2025-12-15T12:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.113087 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.113123 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.113133 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.113148 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.113158 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:20Z","lastTransitionTime":"2025-12-15T12:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.215358 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.215664 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.215674 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.215688 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.215696 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:20Z","lastTransitionTime":"2025-12-15T12:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.318631 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.318673 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.318683 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.318698 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.318711 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:20Z","lastTransitionTime":"2025-12-15T12:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.421980 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.422033 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.422046 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.422064 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.422075 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:20Z","lastTransitionTime":"2025-12-15T12:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.476046 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.476127 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:20 crc kubenswrapper[4719]: E1215 12:18:20.476184 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.476129 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:20 crc kubenswrapper[4719]: E1215 12:18:20.476293 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:20 crc kubenswrapper[4719]: E1215 12:18:20.476423 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.524885 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.524937 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.524951 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.524972 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.524985 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:20Z","lastTransitionTime":"2025-12-15T12:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.626700 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.626770 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.626786 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.626810 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.626825 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:20Z","lastTransitionTime":"2025-12-15T12:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.730240 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.730308 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.730321 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.730347 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.730372 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:20Z","lastTransitionTime":"2025-12-15T12:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.834153 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.834216 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.834238 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.834267 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.834290 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:20Z","lastTransitionTime":"2025-12-15T12:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.936768 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.937165 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.937288 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.937410 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:20 crc kubenswrapper[4719]: I1215 12:18:20.937570 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:20Z","lastTransitionTime":"2025-12-15T12:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.040906 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.040952 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.040968 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.040990 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.041007 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:21Z","lastTransitionTime":"2025-12-15T12:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.143604 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.143639 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.143647 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.143660 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.143668 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:21Z","lastTransitionTime":"2025-12-15T12:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.246998 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.247063 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.247087 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.247116 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.247140 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:21Z","lastTransitionTime":"2025-12-15T12:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.349794 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.350994 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.351195 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.351350 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.351485 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:21Z","lastTransitionTime":"2025-12-15T12:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.454431 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.454485 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.454499 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.454515 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.454527 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:21Z","lastTransitionTime":"2025-12-15T12:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.475957 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:21 crc kubenswrapper[4719]: E1215 12:18:21.476081 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.557451 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.557505 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.557517 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.557532 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.557542 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:21Z","lastTransitionTime":"2025-12-15T12:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.659808 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.660091 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.660168 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.660248 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.660334 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:21Z","lastTransitionTime":"2025-12-15T12:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.762429 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.762680 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.762755 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.762832 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.762956 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:21Z","lastTransitionTime":"2025-12-15T12:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.865402 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.865788 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.866076 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.866287 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.866765 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:21Z","lastTransitionTime":"2025-12-15T12:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.970237 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.970312 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.970335 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.970358 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:21 crc kubenswrapper[4719]: I1215 12:18:21.970375 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:21Z","lastTransitionTime":"2025-12-15T12:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.073324 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.073618 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.073715 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.073840 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.073967 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:22Z","lastTransitionTime":"2025-12-15T12:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.177028 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.177413 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.177604 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.177804 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.178062 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:22Z","lastTransitionTime":"2025-12-15T12:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.284086 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.284134 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.284148 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.284162 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.284170 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:22Z","lastTransitionTime":"2025-12-15T12:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.386628 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.386673 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.386686 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.386712 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.386725 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:22Z","lastTransitionTime":"2025-12-15T12:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.475934 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.475962 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.475948 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:22 crc kubenswrapper[4719]: E1215 12:18:22.476063 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:22 crc kubenswrapper[4719]: E1215 12:18:22.476142 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:22 crc kubenswrapper[4719]: E1215 12:18:22.476214 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.488698 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.488756 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.488770 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.488790 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.488803 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:22Z","lastTransitionTime":"2025-12-15T12:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.591378 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.591418 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.591430 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.591446 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.591461 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:22Z","lastTransitionTime":"2025-12-15T12:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.694065 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.694103 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.694114 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.694129 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.694140 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:22Z","lastTransitionTime":"2025-12-15T12:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.796536 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.796569 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.796578 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.796591 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.796601 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:22Z","lastTransitionTime":"2025-12-15T12:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.898398 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.898436 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.898448 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.898462 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.898473 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:22Z","lastTransitionTime":"2025-12-15T12:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.999892 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.999920 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:22 crc kubenswrapper[4719]: I1215 12:18:22.999931 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:22.999976 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:22.999988 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:22Z","lastTransitionTime":"2025-12-15T12:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.102173 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.102213 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.102237 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.102253 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.102263 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.207315 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.207353 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.207366 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.207386 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.207397 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.309519 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.309552 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.309564 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.309579 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.309590 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.411711 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.411743 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.411752 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.411768 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.411778 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.475948 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:23 crc kubenswrapper[4719]: E1215 12:18:23.476069 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.495212 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.495255 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.495267 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.495282 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.495295 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: E1215 12:18:23.507167 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:23Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.510055 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.510218 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.510307 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.510402 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.510465 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: E1215 12:18:23.521303 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:23Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.524650 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.524783 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.524953 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.525105 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.525257 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: E1215 12:18:23.537046 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:23Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.540193 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.540255 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.540272 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.540289 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.540298 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: E1215 12:18:23.550833 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:23Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.554721 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.555039 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.555121 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.555265 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.555363 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: E1215 12:18:23.565578 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:23Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:23 crc kubenswrapper[4719]: E1215 12:18:23.565747 4719 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.567271 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.567295 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.567303 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.567317 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.567328 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.669762 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.669817 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.669833 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.669877 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.669894 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.772655 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.772705 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.772716 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.772732 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.772742 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.875718 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.875789 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.875802 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.875820 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.875865 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.978527 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.978560 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.978568 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.978584 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:23 crc kubenswrapper[4719]: I1215 12:18:23.978595 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:23Z","lastTransitionTime":"2025-12-15T12:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.081889 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.081932 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.081944 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.081961 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.081973 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:24Z","lastTransitionTime":"2025-12-15T12:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.184669 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.185078 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.185243 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.185403 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.185545 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:24Z","lastTransitionTime":"2025-12-15T12:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.288282 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.288586 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.288675 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.288764 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.288845 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:24Z","lastTransitionTime":"2025-12-15T12:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.390926 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.390955 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.390963 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.390975 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.390984 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:24Z","lastTransitionTime":"2025-12-15T12:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.475611 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.475622 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.475686 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:24 crc kubenswrapper[4719]: E1215 12:18:24.475783 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:24 crc kubenswrapper[4719]: E1215 12:18:24.475890 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:24 crc kubenswrapper[4719]: E1215 12:18:24.476066 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.492661 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.492687 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.492695 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.492706 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.492717 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:24Z","lastTransitionTime":"2025-12-15T12:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.595220 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.595259 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.595269 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.595284 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.595294 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:24Z","lastTransitionTime":"2025-12-15T12:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.698912 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.698949 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.698957 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.698971 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.698981 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:24Z","lastTransitionTime":"2025-12-15T12:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.801480 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.801515 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.801526 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.801560 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.801573 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:24Z","lastTransitionTime":"2025-12-15T12:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.903250 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.903288 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.903297 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.903315 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:24 crc kubenswrapper[4719]: I1215 12:18:24.903324 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:24Z","lastTransitionTime":"2025-12-15T12:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.005934 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.005972 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.005984 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.006000 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.006011 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:25Z","lastTransitionTime":"2025-12-15T12:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.107609 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.107876 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.107967 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.108057 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.108136 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:25Z","lastTransitionTime":"2025-12-15T12:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.210501 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.210695 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.210787 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.210884 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.210964 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:25Z","lastTransitionTime":"2025-12-15T12:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.313239 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.313268 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.313278 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.313294 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.313304 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:25Z","lastTransitionTime":"2025-12-15T12:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.415418 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.415451 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.415462 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.415478 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.415488 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:25Z","lastTransitionTime":"2025-12-15T12:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.476023 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:25 crc kubenswrapper[4719]: E1215 12:18:25.476134 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.519241 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.519271 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.519303 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.519319 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.519329 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:25Z","lastTransitionTime":"2025-12-15T12:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.621556 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.621609 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.621619 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.621632 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.621641 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:25Z","lastTransitionTime":"2025-12-15T12:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.724143 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.724198 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.724210 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.724229 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.724240 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:25Z","lastTransitionTime":"2025-12-15T12:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.826909 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.826941 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.826951 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.826965 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.826977 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:25Z","lastTransitionTime":"2025-12-15T12:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.930284 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.930314 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.930323 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.930339 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:25 crc kubenswrapper[4719]: I1215 12:18:25.930350 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:25Z","lastTransitionTime":"2025-12-15T12:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.032839 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.032886 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.032894 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.032908 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.032917 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:26Z","lastTransitionTime":"2025-12-15T12:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.134528 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.134642 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.134654 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.134671 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.134683 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:26Z","lastTransitionTime":"2025-12-15T12:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.237751 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.238071 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.238144 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.238221 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.238311 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:26Z","lastTransitionTime":"2025-12-15T12:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.340813 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.340900 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.340912 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.340929 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.340940 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:26Z","lastTransitionTime":"2025-12-15T12:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.442804 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.442840 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.442848 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.442878 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.442889 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:26Z","lastTransitionTime":"2025-12-15T12:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.475131 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.475171 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.475159 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.475920 4719 scope.go:117] "RemoveContainer" containerID="a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3" Dec 15 12:18:26 crc kubenswrapper[4719]: E1215 12:18:26.475989 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:26 crc kubenswrapper[4719]: E1215 12:18:26.475976 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:26 crc kubenswrapper[4719]: E1215 12:18:26.475921 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:26 crc kubenswrapper[4719]: E1215 12:18:26.476550 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.545262 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.545307 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.545324 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.545348 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.545364 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:26Z","lastTransitionTime":"2025-12-15T12:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.647803 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.648021 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.648119 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.648207 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.648366 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:26Z","lastTransitionTime":"2025-12-15T12:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.750592 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.750780 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.750880 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.750955 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.751014 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:26Z","lastTransitionTime":"2025-12-15T12:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.853067 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.853104 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.853115 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.853131 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.853141 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:26Z","lastTransitionTime":"2025-12-15T12:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.954968 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.955014 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.955031 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.955050 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:26 crc kubenswrapper[4719]: I1215 12:18:26.955063 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:26Z","lastTransitionTime":"2025-12-15T12:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.056898 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.056933 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.056941 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.056954 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.056963 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:27Z","lastTransitionTime":"2025-12-15T12:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.158759 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.158810 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.158818 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.158831 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.158840 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:27Z","lastTransitionTime":"2025-12-15T12:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.260846 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.260927 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.260939 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.260956 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.260968 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:27Z","lastTransitionTime":"2025-12-15T12:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.363293 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.363324 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.363333 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.363345 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.363353 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:27Z","lastTransitionTime":"2025-12-15T12:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.465498 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.465541 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.465553 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.465569 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.465581 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:27Z","lastTransitionTime":"2025-12-15T12:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.475105 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:27 crc kubenswrapper[4719]: E1215 12:18:27.475274 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.568197 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.568237 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.568247 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.568261 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.568273 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:27Z","lastTransitionTime":"2025-12-15T12:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.670842 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.671122 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.671217 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.671307 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.671398 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:27Z","lastTransitionTime":"2025-12-15T12:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.774387 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.774433 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.774444 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.774461 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.774472 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:27Z","lastTransitionTime":"2025-12-15T12:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.876776 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.877112 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.877203 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.877275 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.877346 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:27Z","lastTransitionTime":"2025-12-15T12:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.980131 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.980179 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.980206 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.980223 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:27 crc kubenswrapper[4719]: I1215 12:18:27.980234 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:27Z","lastTransitionTime":"2025-12-15T12:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.082436 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.082684 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.082755 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.082823 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.082904 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:28Z","lastTransitionTime":"2025-12-15T12:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.185503 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.185550 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.185582 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.185598 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.185608 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:28Z","lastTransitionTime":"2025-12-15T12:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.287753 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.287797 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.287807 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.287823 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.287832 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:28Z","lastTransitionTime":"2025-12-15T12:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.389716 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.389913 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.389929 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.389945 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.389957 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:28Z","lastTransitionTime":"2025-12-15T12:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.672104 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:28 crc kubenswrapper[4719]: E1215 12:18:28.672201 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.672385 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.672431 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:28 crc kubenswrapper[4719]: E1215 12:18:28.672515 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:28 crc kubenswrapper[4719]: E1215 12:18:28.672662 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.673149 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.673173 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.673182 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.673192 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.673201 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:28Z","lastTransitionTime":"2025-12-15T12:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.774969 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.775012 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.775049 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.775064 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.775074 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:28Z","lastTransitionTime":"2025-12-15T12:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.877805 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.877905 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.878171 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.878492 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.878545 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:28Z","lastTransitionTime":"2025-12-15T12:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.981139 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.981383 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.981445 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.981515 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:28 crc kubenswrapper[4719]: I1215 12:18:28.981573 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:28Z","lastTransitionTime":"2025-12-15T12:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.084102 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.084296 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.084382 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.084448 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.084509 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:29Z","lastTransitionTime":"2025-12-15T12:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.186200 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.186307 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.186318 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.186331 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.186340 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:29Z","lastTransitionTime":"2025-12-15T12:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.289324 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.289363 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.289372 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.289386 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.289407 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:29Z","lastTransitionTime":"2025-12-15T12:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.392075 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.392110 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.392120 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.392133 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.392143 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:29Z","lastTransitionTime":"2025-12-15T12:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.476012 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:29 crc kubenswrapper[4719]: E1215 12:18:29.476143 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.489697 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.493929 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.493985 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.493996 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.494009 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.494017 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:29Z","lastTransitionTime":"2025-12-15T12:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.498685 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.510891 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.520529 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.528784 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.537890 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.548996 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.556769 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.568827 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.584061 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.595592 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.595620 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.595628 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.595643 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.595652 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:29Z","lastTransitionTime":"2025-12-15T12:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.596624 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.607110 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.619396 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.630719 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.641897 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.657742 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:12Z\\\",\\\"message\\\":\\\"3047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00720970f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.669730 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:29Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.697138 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.697206 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.697219 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.697238 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.697249 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:29Z","lastTransitionTime":"2025-12-15T12:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.798927 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.798958 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.798969 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.798983 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.798996 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:29Z","lastTransitionTime":"2025-12-15T12:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.881160 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:29 crc kubenswrapper[4719]: E1215 12:18:29.881296 4719 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:18:29 crc kubenswrapper[4719]: E1215 12:18:29.881349 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs podName:0f0b6df0-0612-46c9-83e3-b96f0ec5978d nodeName:}" failed. No retries permitted until 2025-12-15 12:19:01.881335516 +0000 UTC m=+102.823628546 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs") pod "network-metrics-daemon-9kh99" (UID: "0f0b6df0-0612-46c9-83e3-b96f0ec5978d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.900646 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.900681 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.900690 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.900703 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:29 crc kubenswrapper[4719]: I1215 12:18:29.900714 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:29Z","lastTransitionTime":"2025-12-15T12:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.003179 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.003227 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.003237 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.003250 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.003261 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:30Z","lastTransitionTime":"2025-12-15T12:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.105704 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.105738 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.105750 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.105764 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.105774 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:30Z","lastTransitionTime":"2025-12-15T12:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.207735 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.207791 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.207800 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.207812 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.207820 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:30Z","lastTransitionTime":"2025-12-15T12:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.310321 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.310358 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.310372 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.310387 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.310397 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:30Z","lastTransitionTime":"2025-12-15T12:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.413166 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.413202 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.413211 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.413226 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.413235 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:30Z","lastTransitionTime":"2025-12-15T12:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.475601 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.475623 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:30 crc kubenswrapper[4719]: E1215 12:18:30.475713 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.475730 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:30 crc kubenswrapper[4719]: E1215 12:18:30.475828 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:30 crc kubenswrapper[4719]: E1215 12:18:30.475937 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.515845 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.515896 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.515904 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.515919 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.515928 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:30Z","lastTransitionTime":"2025-12-15T12:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.618465 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.618528 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.618542 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.618567 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.618584 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:30Z","lastTransitionTime":"2025-12-15T12:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.721349 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.721388 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.721397 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.721416 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.721425 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:30Z","lastTransitionTime":"2025-12-15T12:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.824782 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.824823 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.824835 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.824869 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.824882 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:30Z","lastTransitionTime":"2025-12-15T12:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.900728 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lcgn4_e63b3458-45ea-47c9-a9ff-7651b7dd4cca/kube-multus/0.log" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.900784 4719 generic.go:334] "Generic (PLEG): container finished" podID="e63b3458-45ea-47c9-a9ff-7651b7dd4cca" containerID="01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae" exitCode=1 Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.900813 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lcgn4" event={"ID":"e63b3458-45ea-47c9-a9ff-7651b7dd4cca","Type":"ContainerDied","Data":"01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.901192 4719 scope.go:117] "RemoveContainer" containerID="01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.913469 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:30Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.925573 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:30Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.928542 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.928582 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.928593 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.928609 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.928620 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:30Z","lastTransitionTime":"2025-12-15T12:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.937379 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:30Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.949334 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:30Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.960479 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:30Z\\\",\\\"message\\\":\\\"2025-12-15T12:17:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9\\\\n2025-12-15T12:17:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9 to /host/opt/cni/bin/\\\\n2025-12-15T12:17:45Z [verbose] multus-daemon started\\\\n2025-12-15T12:17:45Z [verbose] Readiness Indicator file check\\\\n2025-12-15T12:18:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:30Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.970722 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:30Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:30 crc kubenswrapper[4719]: I1215 12:18:30.984588 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:30Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.000846 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:30Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.012305 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.030379 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.030410 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.030419 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.030431 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.030439 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:31Z","lastTransitionTime":"2025-12-15T12:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.034483 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:12Z\\\",\\\"message\\\":\\\"3047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00720970f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.046446 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.060710 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.075220 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.088929 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.103207 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.117361 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.130110 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.132906 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.132933 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.132944 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.132960 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.132970 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:31Z","lastTransitionTime":"2025-12-15T12:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.235508 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.235553 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.235561 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.235576 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.235585 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:31Z","lastTransitionTime":"2025-12-15T12:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.337843 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.337894 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.337903 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.337916 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.337925 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:31Z","lastTransitionTime":"2025-12-15T12:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.440679 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.440715 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.440725 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.440740 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.440750 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:31Z","lastTransitionTime":"2025-12-15T12:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.475667 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:31 crc kubenswrapper[4719]: E1215 12:18:31.475799 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.542528 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.542563 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.542572 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.542585 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.542595 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:31Z","lastTransitionTime":"2025-12-15T12:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.644710 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.644741 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.644756 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.644772 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.644782 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:31Z","lastTransitionTime":"2025-12-15T12:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.747561 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.747603 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.747613 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.747628 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.747637 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:31Z","lastTransitionTime":"2025-12-15T12:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.849626 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.849654 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.849662 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.849675 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.849682 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:31Z","lastTransitionTime":"2025-12-15T12:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.905137 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lcgn4_e63b3458-45ea-47c9-a9ff-7651b7dd4cca/kube-multus/0.log" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.905194 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lcgn4" event={"ID":"e63b3458-45ea-47c9-a9ff-7651b7dd4cca","Type":"ContainerStarted","Data":"64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.918148 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.930456 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.943360 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.951573 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.951618 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.951629 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.951646 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.951657 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:31Z","lastTransitionTime":"2025-12-15T12:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.959679 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:12Z\\\",\\\"message\\\":\\\"3047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00720970f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.971428 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.980795 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:31 crc kubenswrapper[4719]: I1215 12:18:31.993993 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:31Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.007081 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:32Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.016637 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:32Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.028609 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:32Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.041069 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:32Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.051828 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:32Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.053537 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.053575 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.053587 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.053602 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.053614 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:32Z","lastTransitionTime":"2025-12-15T12:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.064644 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:32Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.076098 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:32Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.086217 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:32Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.107071 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:30Z\\\",\\\"message\\\":\\\"2025-12-15T12:17:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9\\\\n2025-12-15T12:17:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9 to /host/opt/cni/bin/\\\\n2025-12-15T12:17:45Z [verbose] multus-daemon started\\\\n2025-12-15T12:17:45Z [verbose] Readiness Indicator file check\\\\n2025-12-15T12:18:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:32Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.120637 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:32Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.163930 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.164159 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.164218 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.164315 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.164372 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:32Z","lastTransitionTime":"2025-12-15T12:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.266917 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.266953 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.266964 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.266981 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.266993 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:32Z","lastTransitionTime":"2025-12-15T12:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.369455 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.369497 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.369505 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.369520 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.369529 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:32Z","lastTransitionTime":"2025-12-15T12:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.471600 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.471637 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.471646 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.471660 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.471669 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:32Z","lastTransitionTime":"2025-12-15T12:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.475873 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.475904 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.475874 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:32 crc kubenswrapper[4719]: E1215 12:18:32.475987 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:32 crc kubenswrapper[4719]: E1215 12:18:32.476355 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:32 crc kubenswrapper[4719]: E1215 12:18:32.476410 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.573740 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.573776 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.573788 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.573806 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.573817 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:32Z","lastTransitionTime":"2025-12-15T12:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.675969 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.676274 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.676376 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.676459 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.676535 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:32Z","lastTransitionTime":"2025-12-15T12:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.779504 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.780097 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.780195 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.780289 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.780370 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:32Z","lastTransitionTime":"2025-12-15T12:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.882069 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.882158 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.882168 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.882183 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.882192 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:32Z","lastTransitionTime":"2025-12-15T12:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.984478 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.984508 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.984516 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.984529 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:32 crc kubenswrapper[4719]: I1215 12:18:32.984537 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:32Z","lastTransitionTime":"2025-12-15T12:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.086893 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.086925 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.086934 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.086947 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.086956 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.189089 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.189124 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.189135 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.189150 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.189162 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.290997 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.291044 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.291087 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.291107 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.291120 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.393480 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.393550 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.393573 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.393605 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.393629 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.475539 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:33 crc kubenswrapper[4719]: E1215 12:18:33.475729 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.496724 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.496781 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.496800 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.496826 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.496845 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.599549 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.599620 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.599636 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.599665 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.599677 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.702226 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.702266 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.702277 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.702293 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.702305 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.806142 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.806516 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.806539 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.806567 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.806585 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.811632 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.811698 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.811719 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.811744 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.811763 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: E1215 12:18:33.829768 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:33Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.833428 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.833464 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.833475 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.833515 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.833527 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: E1215 12:18:33.852508 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:33Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.857481 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.857537 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.857551 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.857567 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.857579 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: E1215 12:18:33.875270 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:33Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.878958 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.879004 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.879016 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.879033 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.879045 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: E1215 12:18:33.892356 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:33Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.895553 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.895582 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.895590 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.895620 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.895630 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:33 crc kubenswrapper[4719]: E1215 12:18:33.911223 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:33Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:33 crc kubenswrapper[4719]: E1215 12:18:33.911378 4719 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.913820 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.913890 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.913903 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.913922 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:33 crc kubenswrapper[4719]: I1215 12:18:33.913937 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:33Z","lastTransitionTime":"2025-12-15T12:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.016605 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.016645 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.016656 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.016670 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.016681 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:34Z","lastTransitionTime":"2025-12-15T12:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.118793 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.118827 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.118836 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.118849 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.118877 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:34Z","lastTransitionTime":"2025-12-15T12:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.221280 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.221336 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.221352 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.221376 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.221393 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:34Z","lastTransitionTime":"2025-12-15T12:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.323622 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.323650 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.323659 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.323673 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.323682 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:34Z","lastTransitionTime":"2025-12-15T12:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.425507 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.425539 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.425547 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.425560 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.425568 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:34Z","lastTransitionTime":"2025-12-15T12:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.475453 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:34 crc kubenswrapper[4719]: E1215 12:18:34.475571 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.475475 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.475463 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:34 crc kubenswrapper[4719]: E1215 12:18:34.475643 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:34 crc kubenswrapper[4719]: E1215 12:18:34.475809 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.527600 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.527641 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.527653 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.527669 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.527678 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:34Z","lastTransitionTime":"2025-12-15T12:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.629719 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.629754 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.629766 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.629782 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.629794 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:34Z","lastTransitionTime":"2025-12-15T12:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.732549 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.732588 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.732599 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.732614 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.732626 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:34Z","lastTransitionTime":"2025-12-15T12:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.834918 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.834954 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.834963 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.834977 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.834986 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:34Z","lastTransitionTime":"2025-12-15T12:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.937967 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.938025 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.938050 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.938085 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:34 crc kubenswrapper[4719]: I1215 12:18:34.938111 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:34Z","lastTransitionTime":"2025-12-15T12:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.041127 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.041162 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.041173 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.041189 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.041199 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:35Z","lastTransitionTime":"2025-12-15T12:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.144314 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.144368 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.144381 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.144397 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.144407 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:35Z","lastTransitionTime":"2025-12-15T12:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.247228 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.247276 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.247289 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.247310 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.247323 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:35Z","lastTransitionTime":"2025-12-15T12:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.349545 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.349584 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.349596 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.349611 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.349620 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:35Z","lastTransitionTime":"2025-12-15T12:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.452044 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.452110 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.452127 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.452151 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.452169 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:35Z","lastTransitionTime":"2025-12-15T12:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.475428 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:35 crc kubenswrapper[4719]: E1215 12:18:35.475694 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.554148 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.554186 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.554197 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.554214 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.554224 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:35Z","lastTransitionTime":"2025-12-15T12:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.657716 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.657784 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.657807 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.657825 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.657838 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:35Z","lastTransitionTime":"2025-12-15T12:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.761015 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.761080 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.761107 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.761140 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.761165 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:35Z","lastTransitionTime":"2025-12-15T12:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.863741 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.863882 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.863910 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.863938 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.863958 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:35Z","lastTransitionTime":"2025-12-15T12:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.967755 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.967809 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.967824 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.967843 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:35 crc kubenswrapper[4719]: I1215 12:18:35.967882 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:35Z","lastTransitionTime":"2025-12-15T12:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.070823 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.070896 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.070913 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.070930 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.070940 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:36Z","lastTransitionTime":"2025-12-15T12:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.173206 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.173286 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.173312 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.173342 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.173364 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:36Z","lastTransitionTime":"2025-12-15T12:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.276321 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.276373 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.276389 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.276436 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.276450 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:36Z","lastTransitionTime":"2025-12-15T12:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.379358 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.379423 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.379445 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.379473 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.379496 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:36Z","lastTransitionTime":"2025-12-15T12:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.475030 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.475081 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.475197 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:36 crc kubenswrapper[4719]: E1215 12:18:36.475399 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:36 crc kubenswrapper[4719]: E1215 12:18:36.475469 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:36 crc kubenswrapper[4719]: E1215 12:18:36.475321 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.483174 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.483227 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.483310 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.483392 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.483402 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:36Z","lastTransitionTime":"2025-12-15T12:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.586539 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.586618 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.586640 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.586664 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.586682 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:36Z","lastTransitionTime":"2025-12-15T12:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.688896 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.688934 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.688942 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.688957 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.688966 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:36Z","lastTransitionTime":"2025-12-15T12:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.791159 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.791197 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.791205 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.791221 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.791231 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:36Z","lastTransitionTime":"2025-12-15T12:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.894020 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.894052 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.894061 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.894074 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.894082 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:36Z","lastTransitionTime":"2025-12-15T12:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.996687 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.996742 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.996760 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.996789 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:36 crc kubenswrapper[4719]: I1215 12:18:36.996809 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:36Z","lastTransitionTime":"2025-12-15T12:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.099261 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.099323 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.099340 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.099365 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.099384 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:37Z","lastTransitionTime":"2025-12-15T12:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.202665 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.202723 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.202792 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.202977 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.202998 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:37Z","lastTransitionTime":"2025-12-15T12:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.307902 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.307966 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.307986 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.308018 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.308041 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:37Z","lastTransitionTime":"2025-12-15T12:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.411569 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.411633 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.411650 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.411677 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.411693 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:37Z","lastTransitionTime":"2025-12-15T12:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.475884 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:37 crc kubenswrapper[4719]: E1215 12:18:37.476116 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.516334 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.516402 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.516419 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.516443 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.516460 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:37Z","lastTransitionTime":"2025-12-15T12:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.619247 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.619289 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.619307 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.619330 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.619346 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:37Z","lastTransitionTime":"2025-12-15T12:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.722274 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.722337 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.722354 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.722377 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.722397 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:37Z","lastTransitionTime":"2025-12-15T12:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.825582 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.825910 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.826012 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.826136 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.826249 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:37Z","lastTransitionTime":"2025-12-15T12:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.928780 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.928819 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.928828 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.928840 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:37 crc kubenswrapper[4719]: I1215 12:18:37.928848 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:37Z","lastTransitionTime":"2025-12-15T12:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.031311 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.031585 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.031678 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.031795 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.031910 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:38Z","lastTransitionTime":"2025-12-15T12:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.135217 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.135280 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.135298 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.135320 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.135338 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:38Z","lastTransitionTime":"2025-12-15T12:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.238045 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.238109 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.238134 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.238158 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.238172 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:38Z","lastTransitionTime":"2025-12-15T12:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.341380 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.341457 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.341473 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.341493 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.341507 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:38Z","lastTransitionTime":"2025-12-15T12:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.444972 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.445422 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.445595 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.445752 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.445926 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:38Z","lastTransitionTime":"2025-12-15T12:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.475095 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.475139 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.475095 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:38 crc kubenswrapper[4719]: E1215 12:18:38.475319 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:38 crc kubenswrapper[4719]: E1215 12:18:38.475412 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:38 crc kubenswrapper[4719]: E1215 12:18:38.475507 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.548906 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.549241 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.549402 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.549553 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.549699 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:38Z","lastTransitionTime":"2025-12-15T12:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.652270 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.652370 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.652391 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.652416 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.652432 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:38Z","lastTransitionTime":"2025-12-15T12:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.754970 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.755026 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.755038 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.755059 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.755075 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:38Z","lastTransitionTime":"2025-12-15T12:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.857574 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.857614 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.857622 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.857636 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.857646 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:38Z","lastTransitionTime":"2025-12-15T12:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.959273 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.959338 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.959348 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.959360 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:38 crc kubenswrapper[4719]: I1215 12:18:38.959369 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:38Z","lastTransitionTime":"2025-12-15T12:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.061826 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.061893 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.061904 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.061918 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.061928 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:39Z","lastTransitionTime":"2025-12-15T12:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.165923 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.165991 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.166010 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.166035 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.166051 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:39Z","lastTransitionTime":"2025-12-15T12:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.268435 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.268769 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.268949 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.269101 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.269247 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:39Z","lastTransitionTime":"2025-12-15T12:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.372475 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.372524 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.372540 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.372563 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.372579 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:39Z","lastTransitionTime":"2025-12-15T12:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.474188 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.474218 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.474225 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.474238 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.474247 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:39Z","lastTransitionTime":"2025-12-15T12:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.476320 4719 scope.go:117] "RemoveContainer" containerID="a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.476575 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:39 crc kubenswrapper[4719]: E1215 12:18:39.476634 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.494375 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.507360 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.524051 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.553451 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:12Z\\\",\\\"message\\\":\\\"3047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00720970f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.574067 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.576419 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.576460 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.576476 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.576497 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.576512 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:39Z","lastTransitionTime":"2025-12-15T12:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.592222 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.603929 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.619668 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.631137 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.641617 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.660309 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:30Z\\\",\\\"message\\\":\\\"2025-12-15T12:17:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9\\\\n2025-12-15T12:17:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9 to /host/opt/cni/bin/\\\\n2025-12-15T12:17:45Z [verbose] multus-daemon started\\\\n2025-12-15T12:17:45Z [verbose] Readiness Indicator file check\\\\n2025-12-15T12:18:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.674561 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.678722 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.678780 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.678803 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.678828 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.678846 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:39Z","lastTransitionTime":"2025-12-15T12:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.692599 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.708174 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.726726 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.741347 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.760630 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.782819 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.782847 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.782868 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.782883 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.782891 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:39Z","lastTransitionTime":"2025-12-15T12:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.884904 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.884979 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.884993 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.885010 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.885045 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:39Z","lastTransitionTime":"2025-12-15T12:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.930954 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/2.log" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.932822 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.933197 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.945350 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.957641 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.970305 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.984002 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.987617 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.987665 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.987686 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.987706 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.987718 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:39Z","lastTransitionTime":"2025-12-15T12:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:39 crc kubenswrapper[4719]: I1215 12:18:39.997738 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:39Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.009655 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.024278 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:30Z\\\",\\\"message\\\":\\\"2025-12-15T12:17:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9\\\\n2025-12-15T12:17:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9 to /host/opt/cni/bin/\\\\n2025-12-15T12:17:45Z [verbose] multus-daemon started\\\\n2025-12-15T12:17:45Z [verbose] Readiness Indicator file check\\\\n2025-12-15T12:18:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.036379 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.058328 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.070884 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.089895 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.089943 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.089958 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.089975 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.089986 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:40Z","lastTransitionTime":"2025-12-15T12:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.093110 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:12Z\\\",\\\"message\\\":\\\"3047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00720970f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:18:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.106069 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.124871 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.136417 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.152831 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.165317 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.176483 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.192676 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.192701 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.192710 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.192724 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.192733 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:40Z","lastTransitionTime":"2025-12-15T12:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.295173 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.295199 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.295207 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.295219 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.295227 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:40Z","lastTransitionTime":"2025-12-15T12:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.397516 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.397548 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.397556 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.397568 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.397577 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:40Z","lastTransitionTime":"2025-12-15T12:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.475848 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:40 crc kubenswrapper[4719]: E1215 12:18:40.475986 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.476140 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:40 crc kubenswrapper[4719]: E1215 12:18:40.476185 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.476282 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:40 crc kubenswrapper[4719]: E1215 12:18:40.476323 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.499660 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.499690 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.499699 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.499711 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.499721 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:40Z","lastTransitionTime":"2025-12-15T12:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.602471 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.602527 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.602544 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.602569 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.602585 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:40Z","lastTransitionTime":"2025-12-15T12:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.705144 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.705179 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.705190 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.705205 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.705216 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:40Z","lastTransitionTime":"2025-12-15T12:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.808422 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.808472 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.808489 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.808514 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.808534 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:40Z","lastTransitionTime":"2025-12-15T12:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.911149 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.911445 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.911608 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.911849 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.912076 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:40Z","lastTransitionTime":"2025-12-15T12:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.938392 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/3.log" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.939736 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/2.log" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.944426 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" exitCode=1 Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.944488 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1"} Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.944563 4719 scope.go:117] "RemoveContainer" containerID="a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.945138 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:18:40 crc kubenswrapper[4719]: E1215 12:18:40.945317 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.963128 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:40 crc kubenswrapper[4719]: I1215 12:18:40.975894 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.000743 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5b74b7e774adcb40923f37c933ffd153d0bc553d58ca2e8e7d554f86dcce4d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:12Z\\\",\\\"message\\\":\\\"3047 4299 0 2025-02-23 05:12:23 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-controller-manager-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:kube-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00720970f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: kube-controller-manager-operator,},ClusterIP:10.217.4.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBala\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:40Z\\\",\\\"message\\\":\\\"s:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1215 12:18:40.476820 6673 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1215 12:18:40.476767 6673 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1215 12:18:40.476827 6673 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1215 12:18:40.476834 6673 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1215 12:18:40.476835 6673 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:40Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.012980 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.020517 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.020562 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.020573 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.020589 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.020602 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:41Z","lastTransitionTime":"2025-12-15T12:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.028262 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.043201 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.054372 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.063635 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.074769 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.083713 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.095764 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.108698 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.118355 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.122245 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.122270 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.122277 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.122290 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.122299 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:41Z","lastTransitionTime":"2025-12-15T12:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.128409 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.139186 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:30Z\\\",\\\"message\\\":\\\"2025-12-15T12:17:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9\\\\n2025-12-15T12:17:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9 to /host/opt/cni/bin/\\\\n2025-12-15T12:17:45Z [verbose] multus-daemon started\\\\n2025-12-15T12:17:45Z [verbose] Readiness Indicator file check\\\\n2025-12-15T12:18:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.148220 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.160596 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.224596 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.224633 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.224642 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.224656 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.224666 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:41Z","lastTransitionTime":"2025-12-15T12:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.328421 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.328486 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.328510 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.328540 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.328564 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:41Z","lastTransitionTime":"2025-12-15T12:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.431170 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.431211 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.431254 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.431275 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.431287 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:41Z","lastTransitionTime":"2025-12-15T12:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.476099 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:41 crc kubenswrapper[4719]: E1215 12:18:41.476272 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.493313 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.534051 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.534095 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.534106 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.534120 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.534130 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:41Z","lastTransitionTime":"2025-12-15T12:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.636449 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.636495 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.636506 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.636521 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.636533 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:41Z","lastTransitionTime":"2025-12-15T12:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.739875 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.739925 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.739937 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.739955 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.740013 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:41Z","lastTransitionTime":"2025-12-15T12:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.842337 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.842425 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.842842 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.842960 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.843255 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:41Z","lastTransitionTime":"2025-12-15T12:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.946253 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.946283 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.946293 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.946308 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.946316 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:41Z","lastTransitionTime":"2025-12-15T12:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.949094 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/3.log" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.952101 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:18:41 crc kubenswrapper[4719]: E1215 12:18:41.952220 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.970783 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65e46fe5560f7eec5d6896d1abf6fac217e29c79ad4d0e0fa93e3ba53f61065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:41 crc kubenswrapper[4719]: I1215 12:18:41.984487 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-w4tb7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02119208-83cb-48f3-b72e-e31a01753830\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f71040f7f75f2afda80f18cc001bef6c5b1f5ad8589bbb028a8441c7c117d6ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d246m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-w4tb7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:41Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.006612 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70777f65-d38a-4352-8e39-a8681d7996f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89315e2f55aa98270dbb4f21f0efaba0a90a3e0b8be53b37b5e291b02f19b641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32fdb594bd33c8c6da94cdf5aecaee0c2c8fee0c50a644c55a56a594715ac3fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01be3862c745bc998424a178c34e6a0d9357ab73dfd5c0e6010893af0f11823a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd7d9c4fd84bc113f9113e0d2f63b3efa25026822759627fde588971405a2962\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4405da428a710333f608a44596c7d41ae77f740112a37ac2666d917827579ba8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1335cb63791ab87047abf22db3da0c70322dc26a0d41c93ad6cdb19d220008f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3c9b7915bb3ff98e76b07f8f4ea7ae50a59ff3da0a8ec8f7fc34bfde8e5d256\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bsd2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dwfzn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.020481 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ae283eb-510a-4419-be95-520ea068a831\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b73421759c210c6eea602f31cc231d9933ad51412ccb10d8f30b197e9a8370b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8022fb51d5892c083efc70c71689f08404f30d935eff2b3389ae992b8a77e36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cjdp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-66nd9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.031371 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9kh99" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:57Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvcvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:57Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9kh99\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.045265 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lcgn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e63b3458-45ea-47c9-a9ff-7651b7dd4cca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:30Z\\\",\\\"message\\\":\\\"2025-12-15T12:17:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9\\\\n2025-12-15T12:17:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8d25f69f-fabc-4d58-bc0b-b9380c5a72c9 to /host/opt/cni/bin/\\\\n2025-12-15T12:17:45Z [verbose] multus-daemon started\\\\n2025-12-15T12:17:45Z [verbose] Readiness Indicator file check\\\\n2025-12-15T12:18:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mlt9t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lcgn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.050086 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.050144 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.050162 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.050187 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.050205 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:42Z","lastTransitionTime":"2025-12-15T12:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.057459 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zbphz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86c27adb-05f9-4c0b-96b3-eb144b6f1c8a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a9dfe0380a4c3a8022ad28a5337a1c5f3c907a19e6bd0c40a860e17a4ce0c39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqgnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zbphz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.073322 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a18fe895-3c85-44eb-b44d-54b25a8cc130\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1215 12:17:32.791616 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1215 12:17:32.793230 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1878641810/tls.crt::/tmp/serving-cert-1878641810/tls.key\\\\\\\"\\\\nI1215 12:17:38.620611 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1215 12:17:38.645547 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1215 12:17:38.645612 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1215 12:17:38.645642 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1215 12:17:38.645651 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1215 12:17:38.653834 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1215 12:17:38.653893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1215 12:17:38.653914 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1215 12:17:38.653935 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1215 12:17:38.653941 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1215 12:17:38.653947 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1215 12:17:38.653953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1215 12:17:38.656793 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.091293 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e19b4f71-1f58-43db-9b0e-d76e740c4700\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2aa7480cf031ff33875e81abe7a5ffd796883ece9f8bf2a75f31e18e19880943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55a1d10705df0d71cf99aa46250af95f10892f10ef125b092c9177644b034ce5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0602b49550f6a9adb45bad226788ee951912a61751ea02de5fcafbc3ac1c168\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.106206 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.117175 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156e10a3096d616af2adab8e006b47279a831de82d2772ee39cd3df2bebb1f63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.128564 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaba47bc-655f-4987-b3c6-b8d49ad5c128\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e47135446157732dd75c4db34a842cc0802582c9c8b73c38f990291370f8eb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rkqnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ff9fm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.147145 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f08ecb19-f273-4df8-bf40-d4ead8b3154f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c3c234548bd0641e81d72bdd4beab790aa1c5f94eddfbf23fb2cdb6c9295da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a1930760367ab0745557ad25571252376c042563fdb3c510f13ba9580c4d869\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efe5a1b418ce4815a63a5395c3b85d42b958d937696657ba65a584bc1e76f0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef55f4727c7c7827314c7852e94810138bf3170faee2cf628c9aa20d7a8aa444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc82de3e459599aae8762fe616cf272d5a1332e6bff0166fb1be7b5108e14765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://260b9864653f37af01256617281e26639ac5aaa77ee05b430a02fc5a7c6aec64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://260b9864653f37af01256617281e26639ac5aaa77ee05b430a02fc5a7c6aec64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e98b0c72c02048d0aaa646ee097dcc29e9a1f0d7b4654428d7f2d28bd3c2c652\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e98b0c72c02048d0aaa646ee097dcc29e9a1f0d7b4654428d7f2d28bd3c2c652\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b0c3035819ad1892b3b8e376d155dcda2427ceb864feef972da53b2866537339\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b0c3035819ad1892b3b8e376d155dcda2427ceb864feef972da53b2866537339\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.152749 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.152784 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.152794 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.152810 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.152821 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:42Z","lastTransitionTime":"2025-12-15T12:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.163831 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab31e0e6-f4e9-4d2e-900f-65371717e1cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b903eb6c106683f903f8508e15ba69d3c8b56fd5851e0f76398a1098d4ef95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746ac7c1cc8984c25eda3f9cc8f89f25c2e64cb446c4d842f384ee2fd7a0f71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e994091cd1000b9db1f0450ccdacb0585c8c00ed7397442acc4e32a7cb0725\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e0730c406ff2653fb29a2ce4e25989ec2994c3bc832531f31289591ed6c2346\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.173979 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.184101 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.199215 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d1afde33-4ec0-4b20-9915-1722f99997fe\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-15T12:18:40Z\\\",\\\"message\\\":\\\"s:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1215 12:18:40.476820 6673 services_controller.go:452] Built service openshift-ingress-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1215 12:18:40.476767 6673 lb_config.go:1031] Cluster endpoints for openshift-kube-controller-manager-operator/metrics for network=default are: map[]\\\\nI1215 12:18:40.476827 6673 services_controller.go:453] Built service openshift-ingress-operator/metrics template LB for network=default: []services.LB{}\\\\nI1215 12:18:40.476834 6673 services_controller.go:454] Service openshift-ingress-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1215 12:18:40.476835 6673 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-15T12:18:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-15T12:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-15T12:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znshh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-15T12:17:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-j8shp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.209883 4719 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-15T12:17:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dae353414db12240dfe28007e3fcad117dfc0294eb0cdc128bca0e41dd554e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5990b6cc80764c9427dbb58968c44293763c20648ad48601e2a7d6703dcc702f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-15T12:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:42Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.258712 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.258751 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.258759 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.258772 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.258780 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:42Z","lastTransitionTime":"2025-12-15T12:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.361127 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.361188 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.361206 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.361229 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.361246 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:42Z","lastTransitionTime":"2025-12-15T12:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.463379 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.463412 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.463420 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.463432 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.463441 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:42Z","lastTransitionTime":"2025-12-15T12:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.475616 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.475699 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.475721 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.475756 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.475851 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.476016 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.566494 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.566542 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.566556 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.566626 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.566641 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:42Z","lastTransitionTime":"2025-12-15T12:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.604346 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.604505 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.604486964 +0000 UTC m=+147.546779994 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.604553 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.604579 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.604600 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.604620 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.604705 4719 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.604739 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.604732241 +0000 UTC m=+147.547025271 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.604831 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.604884 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.604894 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.604932 4719 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.604966 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.604957239 +0000 UTC m=+147.547250259 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.604936 4719 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.605064 4719 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.604904 4719 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.605118 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.605100613 +0000 UTC m=+147.547393683 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:18:42 crc kubenswrapper[4719]: E1215 12:18:42.605167 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.605143945 +0000 UTC m=+147.547437015 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.669531 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.669581 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.669596 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.669618 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.669653 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:42Z","lastTransitionTime":"2025-12-15T12:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.772302 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.772358 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.772373 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.772397 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.772416 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:42Z","lastTransitionTime":"2025-12-15T12:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.874330 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.874366 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.874377 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.874393 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.874404 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:42Z","lastTransitionTime":"2025-12-15T12:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.976181 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.976210 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.976220 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.976236 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:42 crc kubenswrapper[4719]: I1215 12:18:42.976247 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:42Z","lastTransitionTime":"2025-12-15T12:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.078605 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.078640 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.078653 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.078668 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.078678 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:43Z","lastTransitionTime":"2025-12-15T12:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.181383 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.181449 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.181461 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.181484 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.181496 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:43Z","lastTransitionTime":"2025-12-15T12:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.284413 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.284445 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.284456 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.284471 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.284482 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:43Z","lastTransitionTime":"2025-12-15T12:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.386605 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.386672 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.386690 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.386713 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.386731 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:43Z","lastTransitionTime":"2025-12-15T12:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.475669 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:43 crc kubenswrapper[4719]: E1215 12:18:43.475829 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.489537 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.489570 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.489581 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.489595 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.489606 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:43Z","lastTransitionTime":"2025-12-15T12:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.591330 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.591365 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.591376 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.591391 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.591405 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:43Z","lastTransitionTime":"2025-12-15T12:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.693316 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.693354 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.693367 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.693383 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.693394 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:43Z","lastTransitionTime":"2025-12-15T12:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.795175 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.795216 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.795226 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.795242 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.795252 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:43Z","lastTransitionTime":"2025-12-15T12:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.897780 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.897879 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.897901 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.897930 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:43 crc kubenswrapper[4719]: I1215 12:18:43.897945 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:43Z","lastTransitionTime":"2025-12-15T12:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.001091 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.001166 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.001182 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.001202 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.001256 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.082031 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.082077 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.082090 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.082104 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.082115 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: E1215 12:18:44.096557 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.101741 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.101811 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.101827 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.101917 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.101937 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: E1215 12:18:44.113034 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.116989 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.117028 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.117052 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.117075 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.117091 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: E1215 12:18:44.130240 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.133677 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.133705 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.133716 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.133730 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.133741 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: E1215 12:18:44.150374 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.153111 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.153166 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.153176 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.153189 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.153197 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: E1215 12:18:44.164286 4719 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-15T12:18:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"751f1c28-6acd-4965-8e8f-c2d4a0a91bc0\\\",\\\"systemUUID\\\":\\\"7eb9371f-defc-4e63-8dd6-10092c139afd\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-15T12:18:44Z is after 2025-08-24T17:21:41Z" Dec 15 12:18:44 crc kubenswrapper[4719]: E1215 12:18:44.164459 4719 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.165883 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.165934 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.165951 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.165970 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.165984 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.269546 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.269616 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.269638 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.269667 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.269689 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.371494 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.371532 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.371541 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.371555 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.371565 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.473354 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.473394 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.473406 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.473420 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.473431 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.475958 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.476002 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.476053 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:44 crc kubenswrapper[4719]: E1215 12:18:44.476225 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:44 crc kubenswrapper[4719]: E1215 12:18:44.476340 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:44 crc kubenswrapper[4719]: E1215 12:18:44.476394 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.575782 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.575869 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.575883 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.575900 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.575914 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.678693 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.678732 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.678743 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.678762 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.678773 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.780962 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.780995 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.781004 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.781017 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.781025 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.883516 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.883556 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.883564 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.883577 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.883585 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.985286 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.985324 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.985335 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.985351 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:44 crc kubenswrapper[4719]: I1215 12:18:44.985361 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:44Z","lastTransitionTime":"2025-12-15T12:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.087230 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.087264 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.087272 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.087284 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.087294 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:45Z","lastTransitionTime":"2025-12-15T12:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.190146 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.190200 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.190213 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.190232 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.190247 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:45Z","lastTransitionTime":"2025-12-15T12:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.292948 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.292980 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.292988 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.293004 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.293013 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:45Z","lastTransitionTime":"2025-12-15T12:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.395005 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.395049 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.395060 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.395076 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.395100 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:45Z","lastTransitionTime":"2025-12-15T12:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.475660 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:45 crc kubenswrapper[4719]: E1215 12:18:45.475826 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.497465 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.497525 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.497542 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.497563 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.497579 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:45Z","lastTransitionTime":"2025-12-15T12:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.600085 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.600125 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.600134 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.600151 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.600162 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:45Z","lastTransitionTime":"2025-12-15T12:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.702262 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.702300 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.702311 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.702325 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.702335 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:45Z","lastTransitionTime":"2025-12-15T12:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.804920 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.804989 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.805011 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.805043 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.805069 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:45Z","lastTransitionTime":"2025-12-15T12:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.907915 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.907955 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.907964 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.907980 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:45 crc kubenswrapper[4719]: I1215 12:18:45.907989 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:45Z","lastTransitionTime":"2025-12-15T12:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.010288 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.010331 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.010341 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.010357 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.010368 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:46Z","lastTransitionTime":"2025-12-15T12:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.112938 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.112973 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.112983 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.112997 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.113007 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:46Z","lastTransitionTime":"2025-12-15T12:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.218418 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.218476 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.218488 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.218505 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.218517 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:46Z","lastTransitionTime":"2025-12-15T12:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.320643 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.320684 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.320698 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.320715 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.320726 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:46Z","lastTransitionTime":"2025-12-15T12:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.422735 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.422803 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.422855 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.422897 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.422909 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:46Z","lastTransitionTime":"2025-12-15T12:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.475009 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.475017 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.475023 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:46 crc kubenswrapper[4719]: E1215 12:18:46.475189 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:46 crc kubenswrapper[4719]: E1215 12:18:46.475254 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:46 crc kubenswrapper[4719]: E1215 12:18:46.475330 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.525042 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.525136 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.525161 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.525187 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.525205 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:46Z","lastTransitionTime":"2025-12-15T12:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.627740 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.627791 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.627808 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.627826 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.627840 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:46Z","lastTransitionTime":"2025-12-15T12:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.731560 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.731624 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.731646 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.731703 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.731717 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:46Z","lastTransitionTime":"2025-12-15T12:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.837679 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.837722 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.837734 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.837751 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.837770 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:46Z","lastTransitionTime":"2025-12-15T12:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.939394 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.939436 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.939451 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.939470 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:46 crc kubenswrapper[4719]: I1215 12:18:46.939485 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:46Z","lastTransitionTime":"2025-12-15T12:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.042842 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.043040 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.043058 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.043080 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.043094 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:47Z","lastTransitionTime":"2025-12-15T12:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.145821 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.145888 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.145900 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.145917 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.145929 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:47Z","lastTransitionTime":"2025-12-15T12:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.247589 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.247663 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.247695 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.247715 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.247728 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:47Z","lastTransitionTime":"2025-12-15T12:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.350192 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.350229 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.350239 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.350253 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.350262 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:47Z","lastTransitionTime":"2025-12-15T12:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.453154 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.453522 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.453542 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.453568 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.453589 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:47Z","lastTransitionTime":"2025-12-15T12:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.475486 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:47 crc kubenswrapper[4719]: E1215 12:18:47.475642 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.556345 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.556385 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.556397 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.556415 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.556426 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:47Z","lastTransitionTime":"2025-12-15T12:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.659945 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.659988 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.660002 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.660020 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.660031 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:47Z","lastTransitionTime":"2025-12-15T12:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.763253 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.763301 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.763335 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.763354 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.763368 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:47Z","lastTransitionTime":"2025-12-15T12:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.866477 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.866523 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.866532 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.866546 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.866565 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:47Z","lastTransitionTime":"2025-12-15T12:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.968429 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.968486 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.968501 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.968520 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:47 crc kubenswrapper[4719]: I1215 12:18:47.968535 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:47Z","lastTransitionTime":"2025-12-15T12:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.070640 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.070707 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.070718 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.070732 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.070741 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:48Z","lastTransitionTime":"2025-12-15T12:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.173524 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.173565 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.173576 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.173594 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.173607 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:48Z","lastTransitionTime":"2025-12-15T12:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.275935 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.275993 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.276012 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.276037 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.276056 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:48Z","lastTransitionTime":"2025-12-15T12:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.378319 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.378357 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.378367 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.378389 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.378401 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:48Z","lastTransitionTime":"2025-12-15T12:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.475461 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.475461 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:48 crc kubenswrapper[4719]: E1215 12:18:48.475647 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.475504 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:48 crc kubenswrapper[4719]: E1215 12:18:48.475816 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:48 crc kubenswrapper[4719]: E1215 12:18:48.475997 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.481039 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.481130 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.481148 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.481172 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.481189 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:48Z","lastTransitionTime":"2025-12-15T12:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.583507 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.583548 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.583557 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.583571 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.583582 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:48Z","lastTransitionTime":"2025-12-15T12:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.686139 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.686196 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.686212 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.686237 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.686257 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:48Z","lastTransitionTime":"2025-12-15T12:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.789016 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.789086 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.789108 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.789135 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.789157 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:48Z","lastTransitionTime":"2025-12-15T12:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.891890 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.891948 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.891971 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.892004 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.892026 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:48Z","lastTransitionTime":"2025-12-15T12:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.994842 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.994900 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.994911 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.994926 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:48 crc kubenswrapper[4719]: I1215 12:18:48.994937 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:48Z","lastTransitionTime":"2025-12-15T12:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.098217 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.098293 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.098319 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.098348 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.098369 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:49Z","lastTransitionTime":"2025-12-15T12:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.200828 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.200920 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.200941 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.200969 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.200990 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:49Z","lastTransitionTime":"2025-12-15T12:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.303482 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.303518 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.303526 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.303538 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.303548 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:49Z","lastTransitionTime":"2025-12-15T12:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.405633 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.405677 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.405692 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.405711 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.405725 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:49Z","lastTransitionTime":"2025-12-15T12:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.476300 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:49 crc kubenswrapper[4719]: E1215 12:18:49.476749 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.495430 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.514376 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.514420 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.514433 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.514449 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.514461 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:49Z","lastTransitionTime":"2025-12-15T12:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.538343 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-dwfzn" podStartSLOduration=66.538319608 podStartE2EDuration="1m6.538319608s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:49.537949546 +0000 UTC m=+90.480242576" watchObservedRunningTime="2025-12-15 12:18:49.538319608 +0000 UTC m=+90.480612678" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.565482 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-66nd9" podStartSLOduration=66.565463778 podStartE2EDuration="1m6.565463778s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:49.551717237 +0000 UTC m=+90.494010267" watchObservedRunningTime="2025-12-15 12:18:49.565463778 +0000 UTC m=+90.507756808" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.608518 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-w4tb7" podStartSLOduration=67.608487535 podStartE2EDuration="1m7.608487535s" podCreationTimestamp="2025-12-15 12:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:49.591848952 +0000 UTC m=+90.534141982" watchObservedRunningTime="2025-12-15 12:18:49.608487535 +0000 UTC m=+90.550780565" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.608696 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=71.608691642 podStartE2EDuration="1m11.608691642s" podCreationTimestamp="2025-12-15 12:17:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:49.608241897 +0000 UTC m=+90.550534947" watchObservedRunningTime="2025-12-15 12:18:49.608691642 +0000 UTC m=+90.550984672" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.616229 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.616270 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.616281 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.616297 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.616309 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:49Z","lastTransitionTime":"2025-12-15T12:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.660132 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podStartSLOduration=66.660114579 podStartE2EDuration="1m6.660114579s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:49.645928364 +0000 UTC m=+90.588221414" watchObservedRunningTime="2025-12-15 12:18:49.660114579 +0000 UTC m=+90.602407609" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.660677 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-lcgn4" podStartSLOduration=66.660668227 podStartE2EDuration="1m6.660668227s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:49.660112939 +0000 UTC m=+90.602405979" watchObservedRunningTime="2025-12-15 12:18:49.660668227 +0000 UTC m=+90.602961257" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.672030 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zbphz" podStartSLOduration=66.672008879 podStartE2EDuration="1m6.672008879s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:49.671562195 +0000 UTC m=+90.613855225" watchObservedRunningTime="2025-12-15 12:18:49.672008879 +0000 UTC m=+90.614301909" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.689366 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=70.689338744 podStartE2EDuration="1m10.689338744s" podCreationTimestamp="2025-12-15 12:17:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:49.68892123 +0000 UTC m=+90.631214270" watchObservedRunningTime="2025-12-15 12:18:49.689338744 +0000 UTC m=+90.631631774" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.719476 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.719526 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.719539 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.719564 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.719579 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:49Z","lastTransitionTime":"2025-12-15T12:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.760186 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=8.760156122 podStartE2EDuration="8.760156122s" podCreationTimestamp="2025-12-15 12:18:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:49.759541532 +0000 UTC m=+90.701834562" watchObservedRunningTime="2025-12-15 12:18:49.760156122 +0000 UTC m=+90.702449152" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.822839 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.822897 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.822908 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.822925 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.822937 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:49Z","lastTransitionTime":"2025-12-15T12:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.926332 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.927091 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.927101 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.927120 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:49 crc kubenswrapper[4719]: I1215 12:18:49.927132 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:49Z","lastTransitionTime":"2025-12-15T12:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.030819 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.030883 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.030893 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.030909 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.030918 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:50Z","lastTransitionTime":"2025-12-15T12:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.133324 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.133356 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.133364 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.133376 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.133384 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:50Z","lastTransitionTime":"2025-12-15T12:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.236179 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.236209 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.236218 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.236233 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.236242 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:50Z","lastTransitionTime":"2025-12-15T12:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.338407 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.338440 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.338448 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.338461 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.338469 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:50Z","lastTransitionTime":"2025-12-15T12:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.441036 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.441164 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.441181 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.441207 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.441224 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:50Z","lastTransitionTime":"2025-12-15T12:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.475765 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.475798 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.475815 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:50 crc kubenswrapper[4719]: E1215 12:18:50.475972 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:50 crc kubenswrapper[4719]: E1215 12:18:50.476053 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:50 crc kubenswrapper[4719]: E1215 12:18:50.476132 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.543886 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.543924 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.543936 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.543952 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.543963 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:50Z","lastTransitionTime":"2025-12-15T12:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.646725 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.646760 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.646769 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.646782 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.646790 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:50Z","lastTransitionTime":"2025-12-15T12:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.748540 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.748579 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.748589 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.748604 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.748613 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:50Z","lastTransitionTime":"2025-12-15T12:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.851248 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.851300 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.851312 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.851333 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.851347 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:50Z","lastTransitionTime":"2025-12-15T12:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.958594 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.958633 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.958643 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.958658 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:50 crc kubenswrapper[4719]: I1215 12:18:50.958669 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:50Z","lastTransitionTime":"2025-12-15T12:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.060658 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.060689 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.060698 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.060709 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.060718 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:51Z","lastTransitionTime":"2025-12-15T12:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.163469 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.163497 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.163506 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.163520 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.163530 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:51Z","lastTransitionTime":"2025-12-15T12:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.266440 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.266499 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.266510 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.266530 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.266543 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:51Z","lastTransitionTime":"2025-12-15T12:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.369229 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.369264 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.369275 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.369290 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.369300 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:51Z","lastTransitionTime":"2025-12-15T12:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.472122 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.472207 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.472230 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.472256 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.472275 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:51Z","lastTransitionTime":"2025-12-15T12:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.475629 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:51 crc kubenswrapper[4719]: E1215 12:18:51.476271 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.575234 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.575287 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.575323 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.575343 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.575362 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:51Z","lastTransitionTime":"2025-12-15T12:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.677393 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.677430 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.677440 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.677456 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.677467 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:51Z","lastTransitionTime":"2025-12-15T12:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.780358 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.780487 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.780504 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.780606 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.780671 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:51Z","lastTransitionTime":"2025-12-15T12:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.883415 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.883457 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.883469 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.883484 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.883494 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:51Z","lastTransitionTime":"2025-12-15T12:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.985966 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.986017 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.986034 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.986056 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:51 crc kubenswrapper[4719]: I1215 12:18:51.986070 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:51Z","lastTransitionTime":"2025-12-15T12:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.088829 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.088899 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.088913 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.088932 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.088945 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:52Z","lastTransitionTime":"2025-12-15T12:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.192454 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.192542 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.192567 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.192596 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.192621 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:52Z","lastTransitionTime":"2025-12-15T12:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.295705 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.295791 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.295815 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.295848 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.295907 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:52Z","lastTransitionTime":"2025-12-15T12:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.397759 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.397797 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.397805 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.397818 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.397828 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:52Z","lastTransitionTime":"2025-12-15T12:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.475948 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.476001 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:52 crc kubenswrapper[4719]: E1215 12:18:52.476072 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:52 crc kubenswrapper[4719]: E1215 12:18:52.476143 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.476012 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:52 crc kubenswrapper[4719]: E1215 12:18:52.476222 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.476820 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:18:52 crc kubenswrapper[4719]: E1215 12:18:52.477001 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.500242 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.500287 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.500297 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.500313 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.500323 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:52Z","lastTransitionTime":"2025-12-15T12:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.602519 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.602552 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.602561 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.602573 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.602582 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:52Z","lastTransitionTime":"2025-12-15T12:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.705154 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.705214 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.705231 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.705257 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.705273 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:52Z","lastTransitionTime":"2025-12-15T12:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.807423 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.807455 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.807463 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.807476 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.807484 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:52Z","lastTransitionTime":"2025-12-15T12:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.911577 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.911933 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.912079 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.912209 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:52 crc kubenswrapper[4719]: I1215 12:18:52.912342 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:52Z","lastTransitionTime":"2025-12-15T12:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.015147 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.015197 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.015209 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.015228 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.015240 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:53Z","lastTransitionTime":"2025-12-15T12:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.118383 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.118780 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.119042 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.119220 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.119355 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:53Z","lastTransitionTime":"2025-12-15T12:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.223170 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.223246 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.223271 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.223300 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.223323 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:53Z","lastTransitionTime":"2025-12-15T12:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.326257 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.326333 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.326357 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.326385 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.326414 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:53Z","lastTransitionTime":"2025-12-15T12:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.429291 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.429334 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.429345 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.429360 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.429371 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:53Z","lastTransitionTime":"2025-12-15T12:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.475090 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:53 crc kubenswrapper[4719]: E1215 12:18:53.475229 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.535038 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.535122 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.535135 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.535153 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.535163 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:53Z","lastTransitionTime":"2025-12-15T12:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.636817 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.636884 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.636900 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.636918 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.636929 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:53Z","lastTransitionTime":"2025-12-15T12:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.739238 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.739277 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.739285 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.739301 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.739310 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:53Z","lastTransitionTime":"2025-12-15T12:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.842014 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.842051 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.842060 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.842074 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.842084 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:53Z","lastTransitionTime":"2025-12-15T12:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.944888 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.944930 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.944942 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.944960 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:53 crc kubenswrapper[4719]: I1215 12:18:53.944979 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:53Z","lastTransitionTime":"2025-12-15T12:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.047727 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.048325 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.048519 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.048682 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.048834 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:54Z","lastTransitionTime":"2025-12-15T12:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.152189 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.152242 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.152288 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.152305 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.152319 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:54Z","lastTransitionTime":"2025-12-15T12:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.255362 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.255621 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.255757 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.255914 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.256062 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:54Z","lastTransitionTime":"2025-12-15T12:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.359268 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.359929 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.359946 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.359965 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.359994 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:54Z","lastTransitionTime":"2025-12-15T12:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.463221 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.463266 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.463277 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.463292 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.463303 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:54Z","lastTransitionTime":"2025-12-15T12:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.472736 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.472796 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.472818 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.472840 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.472888 4719 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-15T12:18:54Z","lastTransitionTime":"2025-12-15T12:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.475160 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.475193 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:54 crc kubenswrapper[4719]: E1215 12:18:54.475275 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.475304 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:54 crc kubenswrapper[4719]: E1215 12:18:54.475400 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:54 crc kubenswrapper[4719]: E1215 12:18:54.475591 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.528709 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.528688023 podStartE2EDuration="45.528688023s" podCreationTimestamp="2025-12-15 12:18:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:49.772475677 +0000 UTC m=+90.714768697" watchObservedRunningTime="2025-12-15 12:18:54.528688023 +0000 UTC m=+95.470981053" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.529340 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq"] Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.529678 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.533067 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.534010 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.534431 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.535054 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.569804 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=5.569786659 podStartE2EDuration="5.569786659s" podCreationTimestamp="2025-12-15 12:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:54.549615263 +0000 UTC m=+95.491908293" watchObservedRunningTime="2025-12-15 12:18:54.569786659 +0000 UTC m=+95.512079689" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.646696 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-service-ca\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.646748 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.646779 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.646819 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.646838 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.767938 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-service-ca\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.768252 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.768341 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.768470 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.768580 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.768693 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.769022 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-service-ca\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.768499 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.778295 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.786367 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/00ebaba6-eb0e-4e74-81fd-4a59f61b5c81-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-474gq\" (UID: \"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.843048 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" Dec 15 12:18:54 crc kubenswrapper[4719]: I1215 12:18:54.990652 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" event={"ID":"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81","Type":"ContainerStarted","Data":"898441e8a6c0d666c9787b3f379d3f4084f1f252c52b0de7f8a54cd75f0355c8"} Dec 15 12:18:55 crc kubenswrapper[4719]: I1215 12:18:55.475361 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:55 crc kubenswrapper[4719]: E1215 12:18:55.476160 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:55 crc kubenswrapper[4719]: I1215 12:18:55.994960 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" event={"ID":"00ebaba6-eb0e-4e74-81fd-4a59f61b5c81","Type":"ContainerStarted","Data":"e9f8aaa9b6d904133f32d17ed96dd1f712fc538be11a9fa9c50933a253c61fb3"} Dec 15 12:18:56 crc kubenswrapper[4719]: I1215 12:18:56.475770 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:56 crc kubenswrapper[4719]: I1215 12:18:56.475829 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:56 crc kubenswrapper[4719]: E1215 12:18:56.475907 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:56 crc kubenswrapper[4719]: E1215 12:18:56.476013 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:56 crc kubenswrapper[4719]: I1215 12:18:56.476792 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:56 crc kubenswrapper[4719]: E1215 12:18:56.477054 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:57 crc kubenswrapper[4719]: I1215 12:18:57.475452 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:57 crc kubenswrapper[4719]: E1215 12:18:57.475618 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:18:58 crc kubenswrapper[4719]: I1215 12:18:58.475837 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:18:58 crc kubenswrapper[4719]: I1215 12:18:58.475897 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:18:58 crc kubenswrapper[4719]: E1215 12:18:58.475946 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:18:58 crc kubenswrapper[4719]: I1215 12:18:58.475840 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:18:58 crc kubenswrapper[4719]: E1215 12:18:58.476052 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:18:58 crc kubenswrapper[4719]: E1215 12:18:58.476120 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:18:59 crc kubenswrapper[4719]: I1215 12:18:59.475225 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:18:59 crc kubenswrapper[4719]: E1215 12:18:59.476092 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:00 crc kubenswrapper[4719]: I1215 12:19:00.475378 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:00 crc kubenswrapper[4719]: I1215 12:19:00.475461 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:00 crc kubenswrapper[4719]: I1215 12:19:00.475497 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:00 crc kubenswrapper[4719]: E1215 12:19:00.475561 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:00 crc kubenswrapper[4719]: E1215 12:19:00.475700 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:00 crc kubenswrapper[4719]: E1215 12:19:00.475760 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:01 crc kubenswrapper[4719]: I1215 12:19:01.476109 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:01 crc kubenswrapper[4719]: E1215 12:19:01.476261 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:01 crc kubenswrapper[4719]: I1215 12:19:01.934169 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:01 crc kubenswrapper[4719]: E1215 12:19:01.934315 4719 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:19:01 crc kubenswrapper[4719]: E1215 12:19:01.934405 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs podName:0f0b6df0-0612-46c9-83e3-b96f0ec5978d nodeName:}" failed. No retries permitted until 2025-12-15 12:20:05.934388119 +0000 UTC m=+166.876681149 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs") pod "network-metrics-daemon-9kh99" (UID: "0f0b6df0-0612-46c9-83e3-b96f0ec5978d") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 15 12:19:02 crc kubenswrapper[4719]: I1215 12:19:02.475820 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:02 crc kubenswrapper[4719]: I1215 12:19:02.475966 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:02 crc kubenswrapper[4719]: I1215 12:19:02.475820 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:02 crc kubenswrapper[4719]: E1215 12:19:02.475967 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:02 crc kubenswrapper[4719]: E1215 12:19:02.476074 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:02 crc kubenswrapper[4719]: E1215 12:19:02.476157 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:03 crc kubenswrapper[4719]: I1215 12:19:03.476002 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:03 crc kubenswrapper[4719]: E1215 12:19:03.476207 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:04 crc kubenswrapper[4719]: I1215 12:19:04.475148 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:04 crc kubenswrapper[4719]: I1215 12:19:04.475148 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:04 crc kubenswrapper[4719]: E1215 12:19:04.475316 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:04 crc kubenswrapper[4719]: E1215 12:19:04.475371 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:04 crc kubenswrapper[4719]: I1215 12:19:04.475165 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:04 crc kubenswrapper[4719]: E1215 12:19:04.475444 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:05 crc kubenswrapper[4719]: I1215 12:19:05.475797 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:05 crc kubenswrapper[4719]: E1215 12:19:05.477288 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:05 crc kubenswrapper[4719]: I1215 12:19:05.477670 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:19:05 crc kubenswrapper[4719]: E1215 12:19:05.477872 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" Dec 15 12:19:06 crc kubenswrapper[4719]: I1215 12:19:06.475248 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:06 crc kubenswrapper[4719]: I1215 12:19:06.475278 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:06 crc kubenswrapper[4719]: E1215 12:19:06.475532 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:06 crc kubenswrapper[4719]: I1215 12:19:06.475582 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:06 crc kubenswrapper[4719]: E1215 12:19:06.475846 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:06 crc kubenswrapper[4719]: E1215 12:19:06.476187 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:07 crc kubenswrapper[4719]: I1215 12:19:07.476074 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:07 crc kubenswrapper[4719]: E1215 12:19:07.476241 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:08 crc kubenswrapper[4719]: I1215 12:19:08.475811 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:08 crc kubenswrapper[4719]: I1215 12:19:08.475811 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:08 crc kubenswrapper[4719]: I1215 12:19:08.475897 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:08 crc kubenswrapper[4719]: E1215 12:19:08.476234 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:08 crc kubenswrapper[4719]: E1215 12:19:08.476365 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:08 crc kubenswrapper[4719]: E1215 12:19:08.476506 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:09 crc kubenswrapper[4719]: I1215 12:19:09.475762 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:09 crc kubenswrapper[4719]: E1215 12:19:09.476948 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:10 crc kubenswrapper[4719]: I1215 12:19:10.475508 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:10 crc kubenswrapper[4719]: I1215 12:19:10.475580 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:10 crc kubenswrapper[4719]: E1215 12:19:10.475700 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:10 crc kubenswrapper[4719]: I1215 12:19:10.475727 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:10 crc kubenswrapper[4719]: E1215 12:19:10.475895 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:10 crc kubenswrapper[4719]: E1215 12:19:10.476034 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:11 crc kubenswrapper[4719]: I1215 12:19:11.475513 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:11 crc kubenswrapper[4719]: E1215 12:19:11.475669 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:12 crc kubenswrapper[4719]: I1215 12:19:12.475413 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:12 crc kubenswrapper[4719]: E1215 12:19:12.475829 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:12 crc kubenswrapper[4719]: I1215 12:19:12.475518 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:12 crc kubenswrapper[4719]: E1215 12:19:12.476404 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:12 crc kubenswrapper[4719]: I1215 12:19:12.475465 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:12 crc kubenswrapper[4719]: E1215 12:19:12.476663 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:13 crc kubenswrapper[4719]: I1215 12:19:13.475661 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:13 crc kubenswrapper[4719]: E1215 12:19:13.475759 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:14 crc kubenswrapper[4719]: I1215 12:19:14.475537 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:14 crc kubenswrapper[4719]: I1215 12:19:14.475587 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:14 crc kubenswrapper[4719]: I1215 12:19:14.475801 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:14 crc kubenswrapper[4719]: E1215 12:19:14.475913 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:14 crc kubenswrapper[4719]: E1215 12:19:14.475784 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:14 crc kubenswrapper[4719]: E1215 12:19:14.475989 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:15 crc kubenswrapper[4719]: I1215 12:19:15.475788 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:15 crc kubenswrapper[4719]: E1215 12:19:15.475982 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:16 crc kubenswrapper[4719]: I1215 12:19:16.475418 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:16 crc kubenswrapper[4719]: I1215 12:19:16.475416 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:16 crc kubenswrapper[4719]: I1215 12:19:16.475427 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:16 crc kubenswrapper[4719]: E1215 12:19:16.475550 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:16 crc kubenswrapper[4719]: I1215 12:19:16.476282 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:19:16 crc kubenswrapper[4719]: E1215 12:19:16.476326 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:16 crc kubenswrapper[4719]: E1215 12:19:16.476443 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:16 crc kubenswrapper[4719]: E1215 12:19:16.476444 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-j8shp_openshift-ovn-kubernetes(d1afde33-4ec0-4b20-9915-1722f99997fe)\"" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" Dec 15 12:19:17 crc kubenswrapper[4719]: I1215 12:19:17.060563 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lcgn4_e63b3458-45ea-47c9-a9ff-7651b7dd4cca/kube-multus/1.log" Dec 15 12:19:17 crc kubenswrapper[4719]: I1215 12:19:17.061108 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lcgn4_e63b3458-45ea-47c9-a9ff-7651b7dd4cca/kube-multus/0.log" Dec 15 12:19:17 crc kubenswrapper[4719]: I1215 12:19:17.061180 4719 generic.go:334] "Generic (PLEG): container finished" podID="e63b3458-45ea-47c9-a9ff-7651b7dd4cca" containerID="64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427" exitCode=1 Dec 15 12:19:17 crc kubenswrapper[4719]: I1215 12:19:17.061212 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lcgn4" event={"ID":"e63b3458-45ea-47c9-a9ff-7651b7dd4cca","Type":"ContainerDied","Data":"64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427"} Dec 15 12:19:17 crc kubenswrapper[4719]: I1215 12:19:17.061246 4719 scope.go:117] "RemoveContainer" containerID="01d220dda2514994271a2b845c9761a3a0ab828959bedeedefca678d13093eae" Dec 15 12:19:17 crc kubenswrapper[4719]: I1215 12:19:17.061722 4719 scope.go:117] "RemoveContainer" containerID="64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427" Dec 15 12:19:17 crc kubenswrapper[4719]: E1215 12:19:17.063116 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-lcgn4_openshift-multus(e63b3458-45ea-47c9-a9ff-7651b7dd4cca)\"" pod="openshift-multus/multus-lcgn4" podUID="e63b3458-45ea-47c9-a9ff-7651b7dd4cca" Dec 15 12:19:17 crc kubenswrapper[4719]: I1215 12:19:17.090352 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-474gq" podStartSLOduration=94.090324825 podStartE2EDuration="1m34.090324825s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:18:56.007130024 +0000 UTC m=+96.949423084" watchObservedRunningTime="2025-12-15 12:19:17.090324825 +0000 UTC m=+118.032617855" Dec 15 12:19:17 crc kubenswrapper[4719]: I1215 12:19:17.475257 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:17 crc kubenswrapper[4719]: E1215 12:19:17.475387 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:18 crc kubenswrapper[4719]: I1215 12:19:18.065919 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lcgn4_e63b3458-45ea-47c9-a9ff-7651b7dd4cca/kube-multus/1.log" Dec 15 12:19:18 crc kubenswrapper[4719]: I1215 12:19:18.476032 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:18 crc kubenswrapper[4719]: I1215 12:19:18.476172 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:18 crc kubenswrapper[4719]: E1215 12:19:18.476323 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:18 crc kubenswrapper[4719]: E1215 12:19:18.476585 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:18 crc kubenswrapper[4719]: I1215 12:19:18.476815 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:18 crc kubenswrapper[4719]: E1215 12:19:18.477176 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:19 crc kubenswrapper[4719]: E1215 12:19:19.446258 4719 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 15 12:19:19 crc kubenswrapper[4719]: I1215 12:19:19.475548 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:19 crc kubenswrapper[4719]: E1215 12:19:19.477692 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:19 crc kubenswrapper[4719]: E1215 12:19:19.582712 4719 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 15 12:19:20 crc kubenswrapper[4719]: I1215 12:19:20.475004 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:20 crc kubenswrapper[4719]: I1215 12:19:20.475043 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:20 crc kubenswrapper[4719]: I1215 12:19:20.475109 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:20 crc kubenswrapper[4719]: E1215 12:19:20.475114 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:20 crc kubenswrapper[4719]: E1215 12:19:20.475205 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:20 crc kubenswrapper[4719]: E1215 12:19:20.475292 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:21 crc kubenswrapper[4719]: I1215 12:19:21.477046 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:21 crc kubenswrapper[4719]: E1215 12:19:21.477209 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:22 crc kubenswrapper[4719]: I1215 12:19:22.475378 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:22 crc kubenswrapper[4719]: I1215 12:19:22.475419 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:22 crc kubenswrapper[4719]: I1215 12:19:22.475379 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:22 crc kubenswrapper[4719]: E1215 12:19:22.475598 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:22 crc kubenswrapper[4719]: E1215 12:19:22.475734 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:22 crc kubenswrapper[4719]: E1215 12:19:22.475838 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:23 crc kubenswrapper[4719]: I1215 12:19:23.476372 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:23 crc kubenswrapper[4719]: E1215 12:19:23.476517 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:24 crc kubenswrapper[4719]: I1215 12:19:24.475119 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:24 crc kubenswrapper[4719]: I1215 12:19:24.475115 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:24 crc kubenswrapper[4719]: E1215 12:19:24.475473 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:24 crc kubenswrapper[4719]: I1215 12:19:24.475115 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:24 crc kubenswrapper[4719]: E1215 12:19:24.475608 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:24 crc kubenswrapper[4719]: E1215 12:19:24.475914 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:24 crc kubenswrapper[4719]: E1215 12:19:24.584013 4719 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 15 12:19:25 crc kubenswrapper[4719]: I1215 12:19:25.475187 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:25 crc kubenswrapper[4719]: E1215 12:19:25.475332 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:26 crc kubenswrapper[4719]: I1215 12:19:26.475631 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:26 crc kubenswrapper[4719]: I1215 12:19:26.475656 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:26 crc kubenswrapper[4719]: E1215 12:19:26.475784 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:26 crc kubenswrapper[4719]: E1215 12:19:26.475878 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:26 crc kubenswrapper[4719]: I1215 12:19:26.476495 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:26 crc kubenswrapper[4719]: E1215 12:19:26.476656 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:27 crc kubenswrapper[4719]: I1215 12:19:27.475836 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:27 crc kubenswrapper[4719]: E1215 12:19:27.476056 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:28 crc kubenswrapper[4719]: I1215 12:19:28.475114 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:28 crc kubenswrapper[4719]: I1215 12:19:28.475117 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:28 crc kubenswrapper[4719]: I1215 12:19:28.475202 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:28 crc kubenswrapper[4719]: E1215 12:19:28.475413 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:28 crc kubenswrapper[4719]: E1215 12:19:28.475611 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:28 crc kubenswrapper[4719]: E1215 12:19:28.475690 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:28 crc kubenswrapper[4719]: I1215 12:19:28.477185 4719 scope.go:117] "RemoveContainer" containerID="64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427" Dec 15 12:19:28 crc kubenswrapper[4719]: I1215 12:19:28.477341 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:19:29 crc kubenswrapper[4719]: I1215 12:19:29.103931 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/3.log" Dec 15 12:19:29 crc kubenswrapper[4719]: I1215 12:19:29.105971 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerStarted","Data":"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6"} Dec 15 12:19:29 crc kubenswrapper[4719]: I1215 12:19:29.106274 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:19:29 crc kubenswrapper[4719]: I1215 12:19:29.107184 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lcgn4_e63b3458-45ea-47c9-a9ff-7651b7dd4cca/kube-multus/1.log" Dec 15 12:19:29 crc kubenswrapper[4719]: I1215 12:19:29.107208 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lcgn4" event={"ID":"e63b3458-45ea-47c9-a9ff-7651b7dd4cca","Type":"ContainerStarted","Data":"19d7de526bd478b56554fe82bdc23648870759a6be6f40fae4d789f52a7098ee"} Dec 15 12:19:29 crc kubenswrapper[4719]: I1215 12:19:29.168827 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podStartSLOduration=106.168807308 podStartE2EDuration="1m46.168807308s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:29.154661855 +0000 UTC m=+130.096954895" watchObservedRunningTime="2025-12-15 12:19:29.168807308 +0000 UTC m=+130.111100338" Dec 15 12:19:29 crc kubenswrapper[4719]: I1215 12:19:29.479508 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:29 crc kubenswrapper[4719]: E1215 12:19:29.479608 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:29 crc kubenswrapper[4719]: E1215 12:19:29.585831 4719 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 15 12:19:29 crc kubenswrapper[4719]: I1215 12:19:29.685604 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9kh99"] Dec 15 12:19:30 crc kubenswrapper[4719]: I1215 12:19:30.110021 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:30 crc kubenswrapper[4719]: E1215 12:19:30.110391 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:30 crc kubenswrapper[4719]: I1215 12:19:30.475188 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:30 crc kubenswrapper[4719]: E1215 12:19:30.475315 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:30 crc kubenswrapper[4719]: I1215 12:19:30.475383 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:30 crc kubenswrapper[4719]: I1215 12:19:30.475376 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:30 crc kubenswrapper[4719]: E1215 12:19:30.475607 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:30 crc kubenswrapper[4719]: E1215 12:19:30.476072 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:31 crc kubenswrapper[4719]: I1215 12:19:31.475974 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:31 crc kubenswrapper[4719]: E1215 12:19:31.476166 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:32 crc kubenswrapper[4719]: I1215 12:19:32.475180 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:32 crc kubenswrapper[4719]: I1215 12:19:32.475297 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:32 crc kubenswrapper[4719]: E1215 12:19:32.475353 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:32 crc kubenswrapper[4719]: I1215 12:19:32.475315 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:32 crc kubenswrapper[4719]: E1215 12:19:32.475548 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:32 crc kubenswrapper[4719]: E1215 12:19:32.475607 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:33 crc kubenswrapper[4719]: I1215 12:19:33.475318 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:33 crc kubenswrapper[4719]: E1215 12:19:33.475434 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9kh99" podUID="0f0b6df0-0612-46c9-83e3-b96f0ec5978d" Dec 15 12:19:34 crc kubenswrapper[4719]: I1215 12:19:34.475564 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:34 crc kubenswrapper[4719]: E1215 12:19:34.475730 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 15 12:19:34 crc kubenswrapper[4719]: I1215 12:19:34.475795 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:34 crc kubenswrapper[4719]: I1215 12:19:34.475919 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:34 crc kubenswrapper[4719]: E1215 12:19:34.475978 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 15 12:19:34 crc kubenswrapper[4719]: E1215 12:19:34.476131 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.316306 4719 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.361377 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-68lw2"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.362465 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nz9wh"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.362608 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.363199 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.364726 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c5df6"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.365204 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.370294 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.371123 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.372490 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.373434 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.375645 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.376095 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.376588 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.378415 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.379117 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.385697 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.398259 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.410544 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.417401 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.417695 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.418098 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.423590 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.423663 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.423677 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.428500 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.428598 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.428665 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.428776 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.429113 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.429463 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.429562 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.429731 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.429800 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.432333 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.432442 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.432700 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-29pwn"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.433089 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.433475 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.434004 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.434159 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.434280 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.434324 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.434434 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.434522 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.434619 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.435846 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.435988 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.436116 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.436191 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.436306 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.436377 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.436497 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.436570 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.436660 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.436730 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.436828 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.436963 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.437025 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.437084 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.437146 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.438526 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkz2j"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.438964 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.439233 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.439620 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.446019 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-jkk5f"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.446355 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.446556 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7g9lh"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.446822 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.447149 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.447259 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.447351 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.448028 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-2xxrl"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.448294 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.448478 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.448907 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.449803 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9twm9"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.450225 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.450638 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7lswj"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.450913 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.454158 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.454557 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gfmg4"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.454870 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.455083 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.455181 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.455531 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.456704 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.457233 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.457589 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.457796 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.467439 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nz9wh"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.476836 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-glf8c"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.478112 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.478112 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.478414 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.478512 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.478612 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.478669 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.478850 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.478982 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.479244 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.481529 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.481694 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.481969 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.482332 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.482837 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.482883 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.483107 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.483272 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.483441 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.483565 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.483672 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.483699 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.483831 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.483949 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.484068 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.485036 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.485547 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.485651 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.501908 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.502036 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.502179 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.502874 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.503094 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.503532 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.503564 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.503611 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.503885 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.503959 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.504044 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.504135 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-glf8c" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.504210 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.504322 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.504475 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.504580 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.504751 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.506728 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.506845 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.506928 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.507153 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.507530 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.508987 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.509093 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.509350 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.509450 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.509518 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.509632 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-68lw2"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.509984 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.510129 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511038 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511689 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/129630ef-67df-4003-aead-ea250b93c8b5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cdb5p\" (UID: \"129630ef-67df-4003-aead-ea250b93c8b5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511720 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-audit-policies\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511739 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-serving-cert\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511756 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm6cf\" (UniqueName: \"kubernetes.io/projected/129630ef-67df-4003-aead-ea250b93c8b5-kube-api-access-sm6cf\") pod \"openshift-apiserver-operator-796bbdcf4f-cdb5p\" (UID: \"129630ef-67df-4003-aead-ea250b93c8b5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511791 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30f62441-ac5d-48b7-b77c-8b6ff58546d1-serving-cert\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511810 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgf74\" (UniqueName: \"kubernetes.io/projected/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-kube-api-access-jgf74\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511825 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-encryption-config\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511843 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/30f62441-ac5d-48b7-b77c-8b6ff58546d1-etcd-client\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511900 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-audit-dir\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511916 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hp2g\" (UniqueName: \"kubernetes.io/projected/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-kube-api-access-2hp2g\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511932 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/129630ef-67df-4003-aead-ea250b93c8b5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cdb5p\" (UID: \"129630ef-67df-4003-aead-ea250b93c8b5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511946 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-image-import-ca\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511966 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gr5j\" (UniqueName: \"kubernetes.io/projected/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-kube-api-access-9gr5j\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.511983 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-config\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512000 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-config\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512022 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/30f62441-ac5d-48b7-b77c-8b6ff58546d1-encryption-config\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512037 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-etcd-client\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512052 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/30f62441-ac5d-48b7-b77c-8b6ff58546d1-audit-dir\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512069 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512085 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/30f62441-ac5d-48b7-b77c-8b6ff58546d1-node-pullsecrets\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512112 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6665e73-ac68-4fcc-a0e0-85815184a0ea-serving-cert\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512129 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-serving-cert\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512148 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrv8b\" (UniqueName: \"kubernetes.io/projected/e6665e73-ac68-4fcc-a0e0-85815184a0ea-kube-api-access-nrv8b\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512162 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512178 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-audit\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512193 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512207 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0dd6e306-1355-43dd-895f-598d3d873a42-images\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512222 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0dd6e306-1355-43dd-895f-598d3d873a42-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512238 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-client-ca\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512255 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-config\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512270 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqzwv\" (UniqueName: \"kubernetes.io/projected/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-kube-api-access-pqzwv\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512285 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-client-ca\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512306 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-serving-cert\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512330 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-config\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512345 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-etcd-serving-ca\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512361 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28csq\" (UniqueName: \"kubernetes.io/projected/30f62441-ac5d-48b7-b77c-8b6ff58546d1-kube-api-access-28csq\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512377 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512395 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h678z\" (UniqueName: \"kubernetes.io/projected/0dd6e306-1355-43dd-895f-598d3d873a42-kube-api-access-h678z\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512413 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-service-ca-bundle\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512430 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-auth-proxy-config\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512447 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-machine-approver-tls\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512462 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dd6e306-1355-43dd-895f-598d3d873a42-config\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512478 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-config\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.512492 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.513037 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.516976 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.519024 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.519155 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t6qkh"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.519685 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.519727 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.520314 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.522138 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.527978 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.528373 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.528775 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.529442 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.529751 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.529920 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.530082 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.534325 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.535451 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.535827 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.536680 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.538033 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.538685 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.540572 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.540930 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mnl8k"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.541402 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.541640 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.550225 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.575971 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.576082 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.576676 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.577811 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-94r2g"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.579885 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.581063 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.582618 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.599461 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c5df6"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.600955 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.600960 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.601120 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.601488 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.602629 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.609518 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614419 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgf74\" (UniqueName: \"kubernetes.io/projected/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-kube-api-access-jgf74\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614460 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-encryption-config\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614485 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614506 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b44673b6-1b68-46e8-9b5f-b14c3389a017-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-w28lp\" (UID: \"b44673b6-1b68-46e8-9b5f-b14c3389a017\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614533 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/30f62441-ac5d-48b7-b77c-8b6ff58546d1-etcd-client\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614552 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d7d0aab-ec25-437b-99e5-4a39a14bac0c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-p5n7r\" (UID: \"4d7d0aab-ec25-437b-99e5-4a39a14bac0c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614570 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a4fc82e-33f4-4b77-ad47-54554d0c08b4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hxfsm\" (UID: \"3a4fc82e-33f4-4b77-ad47-54554d0c08b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614587 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8f2d8d73-a1cc-4a67-9006-205567f3b907-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fbzqs\" (UID: \"8f2d8d73-a1cc-4a67-9006-205567f3b907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614615 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-audit-dir\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614653 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hp2g\" (UniqueName: \"kubernetes.io/projected/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-kube-api-access-2hp2g\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614668 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-dir\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614689 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/99aced59-b6b7-4abc-8d73-2dc483a14f8f-stats-auth\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614706 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnlxk\" (UniqueName: \"kubernetes.io/projected/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-kube-api-access-jnlxk\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614728 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/129630ef-67df-4003-aead-ea250b93c8b5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cdb5p\" (UID: \"129630ef-67df-4003-aead-ea250b93c8b5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614744 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-image-import-ca\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614762 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gr5j\" (UniqueName: \"kubernetes.io/projected/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-kube-api-access-9gr5j\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614788 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614809 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-config\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614826 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-config\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614873 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/30f62441-ac5d-48b7-b77c-8b6ff58546d1-encryption-config\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614890 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-etcd-client\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614907 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d7d0aab-ec25-437b-99e5-4a39a14bac0c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-p5n7r\" (UID: \"4d7d0aab-ec25-437b-99e5-4a39a14bac0c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614927 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl9rv\" (UniqueName: \"kubernetes.io/projected/3a4fc82e-33f4-4b77-ad47-54554d0c08b4-kube-api-access-nl9rv\") pod \"cluster-samples-operator-665b6dd947-hxfsm\" (UID: \"3a4fc82e-33f4-4b77-ad47-54554d0c08b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614966 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.614986 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/30f62441-ac5d-48b7-b77c-8b6ff58546d1-audit-dir\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615103 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/99aced59-b6b7-4abc-8d73-2dc483a14f8f-metrics-certs\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615145 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615163 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a91fe95-e935-4e7c-bee1-d3d291206f67-trusted-ca\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615180 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f2d8d73-a1cc-4a67-9006-205567f3b907-serving-cert\") pod \"openshift-config-operator-7777fb866f-fbzqs\" (UID: \"8f2d8d73-a1cc-4a67-9006-205567f3b907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615198 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615216 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/30f62441-ac5d-48b7-b77c-8b6ff58546d1-node-pullsecrets\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615233 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615251 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615269 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615306 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6665e73-ac68-4fcc-a0e0-85815184a0ea-serving-cert\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615327 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkrrm\" (UniqueName: \"kubernetes.io/projected/b44673b6-1b68-46e8-9b5f-b14c3389a017-kube-api-access-nkrrm\") pod \"openshift-controller-manager-operator-756b6f6bc6-w28lp\" (UID: \"b44673b6-1b68-46e8-9b5f-b14c3389a017\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615350 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615368 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-serving-cert\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615386 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98dbj\" (UniqueName: \"kubernetes.io/projected/99aced59-b6b7-4abc-8d73-2dc483a14f8f-kube-api-access-98dbj\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615412 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615432 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrv8b\" (UniqueName: \"kubernetes.io/projected/e6665e73-ac68-4fcc-a0e0-85815184a0ea-kube-api-access-nrv8b\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615464 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615480 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-audit\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615499 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615525 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0dd6e306-1355-43dd-895f-598d3d873a42-images\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615543 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0dd6e306-1355-43dd-895f-598d3d873a42-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615557 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-client-ca\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615578 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-config\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.615597 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqzwv\" (UniqueName: \"kubernetes.io/projected/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-kube-api-access-pqzwv\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.627967 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.630946 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.631799 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.632157 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.632981 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/30f62441-ac5d-48b7-b77c-8b6ff58546d1-node-pullsecrets\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.634334 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-config\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.635731 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-image-import-ca\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.637052 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.637675 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.637725 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-audit-dir\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.633231 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/30f62441-ac5d-48b7-b77c-8b6ff58546d1-audit-dir\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.641641 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-config\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.643359 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-client-ca\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.643888 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-audit\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.644339 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0dd6e306-1355-43dd-895f-598d3d873a42-images\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.645813 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.647805 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-trusted-ca-bundle\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648043 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-config\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648565 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6665e73-ac68-4fcc-a0e0-85815184a0ea-serving-cert\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648596 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-client-ca\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648629 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-serving-cert\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648649 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a91fe95-e935-4e7c-bee1-d3d291206f67-config\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648668 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/99aced59-b6b7-4abc-8d73-2dc483a14f8f-default-certificate\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648687 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b03b5a84-a26a-4564-9081-3cfa516c02b3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fr486\" (UID: \"b03b5a84-a26a-4564-9081-3cfa516c02b3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648702 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-policies\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648728 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-config\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648742 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-etcd-serving-ca\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648757 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a91fe95-e935-4e7c-bee1-d3d291206f67-serving-cert\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648774 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28csq\" (UniqueName: \"kubernetes.io/projected/30f62441-ac5d-48b7-b77c-8b6ff58546d1-kube-api-access-28csq\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648791 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648808 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkq28\" (UniqueName: \"kubernetes.io/projected/3a91fe95-e935-4e7c-bee1-d3d291206f67-kube-api-access-zkq28\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648829 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h678z\" (UniqueName: \"kubernetes.io/projected/0dd6e306-1355-43dd-895f-598d3d873a42-kube-api-access-h678z\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648844 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-service-ca-bundle\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648876 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b44673b6-1b68-46e8-9b5f-b14c3389a017-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-w28lp\" (UID: \"b44673b6-1b68-46e8-9b5f-b14c3389a017\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648891 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99aced59-b6b7-4abc-8d73-2dc483a14f8f-service-ca-bundle\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648910 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-auth-proxy-config\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648928 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-machine-approver-tls\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648942 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dd6e306-1355-43dd-895f-598d3d873a42-config\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648960 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-config\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648975 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.648989 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b03b5a84-a26a-4564-9081-3cfa516c02b3-config\") pod \"kube-apiserver-operator-766d6c64bb-fr486\" (UID: \"b03b5a84-a26a-4564-9081-3cfa516c02b3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.649004 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b03b5a84-a26a-4564-9081-3cfa516c02b3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fr486\" (UID: \"b03b5a84-a26a-4564-9081-3cfa516c02b3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.649021 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/129630ef-67df-4003-aead-ea250b93c8b5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cdb5p\" (UID: \"129630ef-67df-4003-aead-ea250b93c8b5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.649035 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-audit-policies\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.649052 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-serving-cert\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.649069 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f27r5\" (UniqueName: \"kubernetes.io/projected/8f2d8d73-a1cc-4a67-9006-205567f3b907-kube-api-access-f27r5\") pod \"openshift-config-operator-7777fb866f-fbzqs\" (UID: \"8f2d8d73-a1cc-4a67-9006-205567f3b907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.649085 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.649104 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm6cf\" (UniqueName: \"kubernetes.io/projected/129630ef-67df-4003-aead-ea250b93c8b5-kube-api-access-sm6cf\") pod \"openshift-apiserver-operator-796bbdcf4f-cdb5p\" (UID: \"129630ef-67df-4003-aead-ea250b93c8b5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.649121 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d7d0aab-ec25-437b-99e5-4a39a14bac0c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-p5n7r\" (UID: \"4d7d0aab-ec25-437b-99e5-4a39a14bac0c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.649150 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30f62441-ac5d-48b7-b77c-8b6ff58546d1-serving-cert\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.650134 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-client-ca\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.650390 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-encryption-config\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.650677 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.651342 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-auth-proxy-config\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.651423 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/30f62441-ac5d-48b7-b77c-8b6ff58546d1-encryption-config\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.651761 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/30f62441-ac5d-48b7-b77c-8b6ff58546d1-etcd-client\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.652102 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/129630ef-67df-4003-aead-ea250b93c8b5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cdb5p\" (UID: \"129630ef-67df-4003-aead-ea250b93c8b5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.652513 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dd6e306-1355-43dd-895f-598d3d873a42-config\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.652974 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-etcd-serving-ca\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.653525 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30f62441-ac5d-48b7-b77c-8b6ff58546d1-config\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.654254 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-machine-approver-tls\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.654279 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.655081 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.655240 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-config\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.655603 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-audit-policies\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.655993 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-service-ca-bundle\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.656790 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-serving-cert\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.656969 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-serving-cert\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.657109 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/129630ef-67df-4003-aead-ea250b93c8b5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cdb5p\" (UID: \"129630ef-67df-4003-aead-ea250b93c8b5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.657437 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0dd6e306-1355-43dd-895f-598d3d873a42-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.658562 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-serving-cert\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.661240 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/30f62441-ac5d-48b7-b77c-8b6ff58546d1-serving-cert\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.663081 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7g9lh"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.665833 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.666842 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-vchb6"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.667358 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-vchb6" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.672587 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2xxrl"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.686294 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-etcd-client\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.687568 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.688671 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-glf8c"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.690524 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.691277 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.696194 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9twm9"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.698899 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.700184 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.701923 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.703702 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.703732 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gfmg4"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.705154 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkz2j"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.705371 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.705939 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.706288 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-29pwn"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.707676 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7lswj"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.708544 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.709903 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t6qkh"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.710875 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.712932 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.712968 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.714332 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.714920 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.717376 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-94r2g"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.717400 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.718828 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.718874 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.720102 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-827sb"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.721013 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-ls926"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.721287 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-827sb" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.721402 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.722463 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.723618 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.724566 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-827sb"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.725544 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.725675 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.726621 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.727753 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-vchb6"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.729478 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4mbgz"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.731074 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mnl8k"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.731162 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.732090 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4mbgz"] Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.745654 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.750970 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-dir\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751010 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/99aced59-b6b7-4abc-8d73-2dc483a14f8f-stats-auth\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751038 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-dir\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751036 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnlxk\" (UniqueName: \"kubernetes.io/projected/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-kube-api-access-jnlxk\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751112 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751155 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d7d0aab-ec25-437b-99e5-4a39a14bac0c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-p5n7r\" (UID: \"4d7d0aab-ec25-437b-99e5-4a39a14bac0c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751179 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl9rv\" (UniqueName: \"kubernetes.io/projected/3a4fc82e-33f4-4b77-ad47-54554d0c08b4-kube-api-access-nl9rv\") pod \"cluster-samples-operator-665b6dd947-hxfsm\" (UID: \"3a4fc82e-33f4-4b77-ad47-54554d0c08b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751204 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751246 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/99aced59-b6b7-4abc-8d73-2dc483a14f8f-metrics-certs\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751268 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f2d8d73-a1cc-4a67-9006-205567f3b907-serving-cert\") pod \"openshift-config-operator-7777fb866f-fbzqs\" (UID: \"8f2d8d73-a1cc-4a67-9006-205567f3b907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751350 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751388 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a91fe95-e935-4e7c-bee1-d3d291206f67-trusted-ca\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751410 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751431 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751450 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751549 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkrrm\" (UniqueName: \"kubernetes.io/projected/b44673b6-1b68-46e8-9b5f-b14c3389a017-kube-api-access-nkrrm\") pod \"openshift-controller-manager-operator-756b6f6bc6-w28lp\" (UID: \"b44673b6-1b68-46e8-9b5f-b14c3389a017\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751594 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751620 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.751646 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98dbj\" (UniqueName: \"kubernetes.io/projected/99aced59-b6b7-4abc-8d73-2dc483a14f8f-kube-api-access-98dbj\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.753047 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.753077 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a91fe95-e935-4e7c-bee1-d3d291206f67-config\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.753387 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/99aced59-b6b7-4abc-8d73-2dc483a14f8f-default-certificate\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.753419 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b03b5a84-a26a-4564-9081-3cfa516c02b3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fr486\" (UID: \"b03b5a84-a26a-4564-9081-3cfa516c02b3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.753511 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.753911 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-policies\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.752934 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a91fe95-e935-4e7c-bee1-d3d291206f67-trusted-ca\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.755921 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.755954 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-policies\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.755977 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a91fe95-e935-4e7c-bee1-d3d291206f67-serving-cert\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756012 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkq28\" (UniqueName: \"kubernetes.io/projected/3a91fe95-e935-4e7c-bee1-d3d291206f67-kube-api-access-zkq28\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756039 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b44673b6-1b68-46e8-9b5f-b14c3389a017-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-w28lp\" (UID: \"b44673b6-1b68-46e8-9b5f-b14c3389a017\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756056 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99aced59-b6b7-4abc-8d73-2dc483a14f8f-service-ca-bundle\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756076 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b03b5a84-a26a-4564-9081-3cfa516c02b3-config\") pod \"kube-apiserver-operator-766d6c64bb-fr486\" (UID: \"b03b5a84-a26a-4564-9081-3cfa516c02b3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756093 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b03b5a84-a26a-4564-9081-3cfa516c02b3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fr486\" (UID: \"b03b5a84-a26a-4564-9081-3cfa516c02b3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756151 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f27r5\" (UniqueName: \"kubernetes.io/projected/8f2d8d73-a1cc-4a67-9006-205567f3b907-kube-api-access-f27r5\") pod \"openshift-config-operator-7777fb866f-fbzqs\" (UID: \"8f2d8d73-a1cc-4a67-9006-205567f3b907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756171 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756198 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d7d0aab-ec25-437b-99e5-4a39a14bac0c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-p5n7r\" (UID: \"4d7d0aab-ec25-437b-99e5-4a39a14bac0c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756229 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756245 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b44673b6-1b68-46e8-9b5f-b14c3389a017-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-w28lp\" (UID: \"b44673b6-1b68-46e8-9b5f-b14c3389a017\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756324 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d7d0aab-ec25-437b-99e5-4a39a14bac0c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-p5n7r\" (UID: \"4d7d0aab-ec25-437b-99e5-4a39a14bac0c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756371 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a4fc82e-33f4-4b77-ad47-54554d0c08b4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hxfsm\" (UID: \"3a4fc82e-33f4-4b77-ad47-54554d0c08b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756400 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8f2d8d73-a1cc-4a67-9006-205567f3b907-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fbzqs\" (UID: \"8f2d8d73-a1cc-4a67-9006-205567f3b907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756677 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756756 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756814 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8f2d8d73-a1cc-4a67-9006-205567f3b907-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fbzqs\" (UID: \"8f2d8d73-a1cc-4a67-9006-205567f3b907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.756987 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/99aced59-b6b7-4abc-8d73-2dc483a14f8f-metrics-certs\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.757063 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b44673b6-1b68-46e8-9b5f-b14c3389a017-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-w28lp\" (UID: \"b44673b6-1b68-46e8-9b5f-b14c3389a017\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.757146 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/99aced59-b6b7-4abc-8d73-2dc483a14f8f-default-certificate\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.757390 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.757637 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99aced59-b6b7-4abc-8d73-2dc483a14f8f-service-ca-bundle\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.757987 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.758005 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d7d0aab-ec25-437b-99e5-4a39a14bac0c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-p5n7r\" (UID: \"4d7d0aab-ec25-437b-99e5-4a39a14bac0c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.758178 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b03b5a84-a26a-4564-9081-3cfa516c02b3-config\") pod \"kube-apiserver-operator-766d6c64bb-fr486\" (UID: \"b03b5a84-a26a-4564-9081-3cfa516c02b3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.758678 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.758811 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/99aced59-b6b7-4abc-8d73-2dc483a14f8f-stats-auth\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.758963 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.759157 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.760243 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a91fe95-e935-4e7c-bee1-d3d291206f67-serving-cert\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.760350 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.760956 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.761979 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d7d0aab-ec25-437b-99e5-4a39a14bac0c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-p5n7r\" (UID: \"4d7d0aab-ec25-437b-99e5-4a39a14bac0c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.762594 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b03b5a84-a26a-4564-9081-3cfa516c02b3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fr486\" (UID: \"b03b5a84-a26a-4564-9081-3cfa516c02b3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.763169 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b44673b6-1b68-46e8-9b5f-b14c3389a017-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-w28lp\" (UID: \"b44673b6-1b68-46e8-9b5f-b14c3389a017\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.764241 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.784938 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.786747 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a91fe95-e935-4e7c-bee1-d3d291206f67-config\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.808623 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.824905 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.845902 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.864366 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.885242 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.905053 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.924363 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.944318 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.964283 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.986000 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 15 12:19:35 crc kubenswrapper[4719]: I1215 12:19:35.994834 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f2d8d73-a1cc-4a67-9006-205567f3b907-serving-cert\") pod \"openshift-config-operator-7777fb866f-fbzqs\" (UID: \"8f2d8d73-a1cc-4a67-9006-205567f3b907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.005429 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.025454 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.044342 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.064984 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.085323 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.105158 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.111406 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a4fc82e-33f4-4b77-ad47-54554d0c08b4-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hxfsm\" (UID: \"3a4fc82e-33f4-4b77-ad47-54554d0c08b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.125296 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.145556 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.164331 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.185263 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.205156 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.245885 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.265443 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.284931 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.308992 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.324965 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.345099 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.366353 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.384751 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.404684 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.424994 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.445574 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.465622 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.475398 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.475423 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.475410 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.485425 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.505735 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.525817 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.543602 4719 request.go:700] Waited for 1.013185494s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmco-proxy-tls&limit=500&resourceVersion=0 Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.545628 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.566282 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.585598 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.605636 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.624829 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.646084 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.664609 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.684950 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.704664 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.724941 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.753306 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.765915 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.786779 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.806173 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.825807 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.866018 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.885630 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.905978 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.926080 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.946171 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.965721 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 15 12:19:36 crc kubenswrapper[4719]: I1215 12:19:36.985365 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.006495 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.025520 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.045306 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.065789 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.086217 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.132765 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgf74\" (UniqueName: \"kubernetes.io/projected/78ebc618-1e6f-47e9-9afe-b25bf76a19ff-kube-api-access-jgf74\") pod \"machine-approver-56656f9798-fszmm\" (UID: \"78ebc618-1e6f-47e9-9afe-b25bf76a19ff\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.134329 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.144920 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.165502 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.186040 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.205234 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.239451 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gr5j\" (UniqueName: \"kubernetes.io/projected/bc81baea-ae9e-4c9d-af05-d15fd4b1591c-kube-api-access-9gr5j\") pod \"authentication-operator-69f744f599-29pwn\" (UID: \"bc81baea-ae9e-4c9d-af05-d15fd4b1591c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.240745 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.264761 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hp2g\" (UniqueName: \"kubernetes.io/projected/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-kube-api-access-2hp2g\") pod \"route-controller-manager-6576b87f9c-sm6zr\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.281205 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrv8b\" (UniqueName: \"kubernetes.io/projected/e6665e73-ac68-4fcc-a0e0-85815184a0ea-kube-api-access-nrv8b\") pod \"controller-manager-879f6c89f-nz9wh\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.295392 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.300008 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqzwv\" (UniqueName: \"kubernetes.io/projected/8b58d079-d1d0-4923-8ba3-bfeb1dc30b72-kube-api-access-pqzwv\") pod \"apiserver-7bbb656c7d-gp49f\" (UID: \"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.320533 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28csq\" (UniqueName: \"kubernetes.io/projected/30f62441-ac5d-48b7-b77c-8b6ff58546d1-kube-api-access-28csq\") pod \"apiserver-76f77b778f-68lw2\" (UID: \"30f62441-ac5d-48b7-b77c-8b6ff58546d1\") " pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.340969 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h678z\" (UniqueName: \"kubernetes.io/projected/0dd6e306-1355-43dd-895f-598d3d873a42-kube-api-access-h678z\") pod \"machine-api-operator-5694c8668f-c5df6\" (UID: \"0dd6e306-1355-43dd-895f-598d3d873a42\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.360852 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm6cf\" (UniqueName: \"kubernetes.io/projected/129630ef-67df-4003-aead-ea250b93c8b5-kube-api-access-sm6cf\") pod \"openshift-apiserver-operator-796bbdcf4f-cdb5p\" (UID: \"129630ef-67df-4003-aead-ea250b93c8b5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.365417 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.385015 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.405610 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.426848 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.445298 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.466253 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.484803 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.492074 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.501161 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.504965 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.509037 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.525391 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.537360 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.543976 4719 request.go:700] Waited for 1.822338652s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-tls&limit=500&resourceVersion=0 Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.545833 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.553514 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.567273 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.586956 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.595280 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.604520 4719 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.651880 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnlxk\" (UniqueName: \"kubernetes.io/projected/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-kube-api-access-jnlxk\") pod \"oauth-openshift-558db77b4-nkz2j\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.664533 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d7d0aab-ec25-437b-99e5-4a39a14bac0c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-p5n7r\" (UID: \"4d7d0aab-ec25-437b-99e5-4a39a14bac0c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.686329 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl9rv\" (UniqueName: \"kubernetes.io/projected/3a4fc82e-33f4-4b77-ad47-54554d0c08b4-kube-api-access-nl9rv\") pod \"cluster-samples-operator-665b6dd947-hxfsm\" (UID: \"3a4fc82e-33f4-4b77-ad47-54554d0c08b4\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.705578 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.717926 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98dbj\" (UniqueName: \"kubernetes.io/projected/99aced59-b6b7-4abc-8d73-2dc483a14f8f-kube-api-access-98dbj\") pod \"router-default-5444994796-jkk5f\" (UID: \"99aced59-b6b7-4abc-8d73-2dc483a14f8f\") " pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.727265 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkrrm\" (UniqueName: \"kubernetes.io/projected/b44673b6-1b68-46e8-9b5f-b14c3389a017-kube-api-access-nkrrm\") pod \"openshift-controller-manager-operator-756b6f6bc6-w28lp\" (UID: \"b44673b6-1b68-46e8-9b5f-b14c3389a017\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.765627 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkq28\" (UniqueName: \"kubernetes.io/projected/3a91fe95-e935-4e7c-bee1-d3d291206f67-kube-api-access-zkq28\") pod \"console-operator-58897d9998-7lswj\" (UID: \"3a91fe95-e935-4e7c-bee1-d3d291206f67\") " pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.778304 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.788575 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f27r5\" (UniqueName: \"kubernetes.io/projected/8f2d8d73-a1cc-4a67-9006-205567f3b907-kube-api-access-f27r5\") pod \"openshift-config-operator-7777fb866f-fbzqs\" (UID: \"8f2d8d73-a1cc-4a67-9006-205567f3b907\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.804843 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.813024 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.825029 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.833562 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.845628 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 15 12:19:37 crc kubenswrapper[4719]: I1215 12:19:37.865525 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.611663 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b03b5a84-a26a-4564-9081-3cfa516c02b3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fr486\" (UID: \"b03b5a84-a26a-4564-9081-3cfa516c02b3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.611664 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.615618 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.618454 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.622416 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-tls\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.623707 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/27255302-b4b6-40dc-a5fd-f6e376b7e05d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.623755 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/27255302-b4b6-40dc-a5fd-f6e376b7e05d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.627544 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.629137 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" event={"ID":"78ebc618-1e6f-47e9-9afe-b25bf76a19ff","Type":"ContainerStarted","Data":"fdf59bd9d0e19df2864c868815405bca0fc63254ec4dca062856d303a8c2d3b8"} Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.629922 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.629976 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-certificates\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.630035 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-bound-sa-token\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.630075 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-trusted-ca\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.630095 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dqww\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-kube-api-access-6dqww\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: E1215 12:19:38.630976 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.130964333 +0000 UTC m=+140.073257363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.676479 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nz9wh"] Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.688368 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-29pwn"] Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.706023 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr"] Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.712680 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c5df6"] Dec 15 12:19:38 crc kubenswrapper[4719]: W1215 12:19:38.712984 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6665e73_ac68_4fcc_a0e0_85815184a0ea.slice/crio-65a7c1c640aced6bbb321c169e97f419b80ceaa9fe61ae6c575ca7f7e61bed92 WatchSource:0}: Error finding container 65a7c1c640aced6bbb321c169e97f419b80ceaa9fe61ae6c575ca7f7e61bed92: Status 404 returned error can't find the container with id 65a7c1c640aced6bbb321c169e97f419b80ceaa9fe61ae6c575ca7f7e61bed92 Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.715301 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-68lw2"] Dec 15 12:19:38 crc kubenswrapper[4719]: W1215 12:19:38.728590 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe9cf63_1287_418f_b2c5_4e4e7458cf44.slice/crio-aae65cef2246ccafec9cf563c241796d847fae667522399f3442c503e8aeef4c WatchSource:0}: Error finding container aae65cef2246ccafec9cf563c241796d847fae667522399f3442c503e8aeef4c: Status 404 returned error can't find the container with id aae65cef2246ccafec9cf563c241796d847fae667522399f3442c503e8aeef4c Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.737563 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.737832 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1-metrics-tls\") pod \"dns-operator-744455d44c-9twm9\" (UID: \"e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1\") " pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" Dec 15 12:19:38 crc kubenswrapper[4719]: E1215 12:19:38.737896 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.237849534 +0000 UTC m=+140.180142564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.737921 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fd3ac7a-862d-45ff-a931-5962b105ac6e-serving-cert\") pod \"service-ca-operator-777779d784-lsvbn\" (UID: \"4fd3ac7a-862d-45ff-a931-5962b105ac6e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.737967 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4cec2f60-1603-4a04-baab-3d9c070257da-etcd-ca\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.737985 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e-proxy-tls\") pod \"machine-config-controller-84d6567774-zhwtq\" (UID: \"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.738002 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zhwtq\" (UID: \"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.738035 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cec2f60-1603-4a04-baab-3d9c070257da-config\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.738058 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-bound-sa-token\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.738074 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-trusted-ca\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.738089 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dqww\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-kube-api-access-6dqww\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.738130 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v2sq\" (UniqueName: \"kubernetes.io/projected/aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e-kube-api-access-4v2sq\") pod \"machine-config-controller-84d6567774-zhwtq\" (UID: \"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.738495 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbszz\" (UniqueName: \"kubernetes.io/projected/4cec2f60-1603-4a04-baab-3d9c070257da-kube-api-access-dbszz\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.739562 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2558ee2-4e5f-419a-94ca-5b37d181f83c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zcgr5\" (UID: \"f2558ee2-4e5f-419a-94ca-5b37d181f83c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.739593 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-oauth-config\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.740705 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4cec2f60-1603-4a04-baab-3d9c070257da-etcd-service-ca\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.741134 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb7kh\" (UniqueName: \"kubernetes.io/projected/556c5cc5-d57e-413d-aa79-421d196cb799-kube-api-access-fb7kh\") pod \"marketplace-operator-79b997595-t6qkh\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.741305 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd3ac7a-862d-45ff-a931-5962b105ac6e-config\") pod \"service-ca-operator-777779d784-lsvbn\" (UID: \"4fd3ac7a-862d-45ff-a931-5962b105ac6e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.741435 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/994235ed-c7e5-44d3-bef7-c8df6b3224a3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.741627 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9rhq\" (UniqueName: \"kubernetes.io/projected/790b2072-3dd0-4e6c-896c-724a3cd2bc5d-kube-api-access-r9rhq\") pod \"package-server-manager-789f6589d5-8wqj5\" (UID: \"790b2072-3dd0-4e6c-896c-724a3cd2bc5d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.741926 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/27255302-b4b6-40dc-a5fd-f6e376b7e05d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.741957 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-serving-cert\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.741985 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-trusted-ca-bundle\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742035 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp24b\" (UniqueName: \"kubernetes.io/projected/e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1-kube-api-access-hp24b\") pod \"dns-operator-744455d44c-9twm9\" (UID: \"e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1\") " pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742057 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742082 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/994235ed-c7e5-44d3-bef7-c8df6b3224a3-proxy-tls\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742103 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a2ade101-54a5-4915-9b0c-bc668fe625ff-srv-cert\") pod \"olm-operator-6b444d44fb-xqsdv\" (UID: \"a2ade101-54a5-4915-9b0c-bc668fe625ff\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742134 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742169 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-certificates\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742194 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t6qkh\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742216 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/994235ed-c7e5-44d3-bef7-c8df6b3224a3-images\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742239 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d620f9fd-1374-4f93-972d-93f2b9173471-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ngv2q\" (UID: \"d620f9fd-1374-4f93-972d-93f2b9173471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742285 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w4mc\" (UniqueName: \"kubernetes.io/projected/4fd3ac7a-862d-45ff-a931-5962b105ac6e-kube-api-access-8w4mc\") pod \"service-ca-operator-777779d784-lsvbn\" (UID: \"4fd3ac7a-862d-45ff-a931-5962b105ac6e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742303 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a2ade101-54a5-4915-9b0c-bc668fe625ff-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xqsdv\" (UID: \"a2ade101-54a5-4915-9b0c-bc668fe625ff\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742323 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26xtv\" (UniqueName: \"kubernetes.io/projected/994235ed-c7e5-44d3-bef7-c8df6b3224a3-kube-api-access-26xtv\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742343 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-service-ca\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742362 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrffv\" (UniqueName: \"kubernetes.io/projected/c8458395-1102-446b-8b75-752820a9cfdb-kube-api-access-vrffv\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742395 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v4tl\" (UniqueName: \"kubernetes.io/projected/d620f9fd-1374-4f93-972d-93f2b9173471-kube-api-access-2v4tl\") pod \"control-plane-machine-set-operator-78cbb6b69f-ngv2q\" (UID: \"d620f9fd-1374-4f93-972d-93f2b9173471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742418 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/790b2072-3dd0-4e6c-896c-724a3cd2bc5d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8wqj5\" (UID: \"790b2072-3dd0-4e6c-896c-724a3cd2bc5d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742478 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4cec2f60-1603-4a04-baab-3d9c070257da-etcd-client\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742501 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2558ee2-4e5f-419a-94ca-5b37d181f83c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zcgr5\" (UID: \"f2558ee2-4e5f-419a-94ca-5b37d181f83c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742541 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4rwv\" (UniqueName: \"kubernetes.io/projected/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-kube-api-access-d4rwv\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742571 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-tls\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742596 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcdbd\" (UniqueName: \"kubernetes.io/projected/a2ade101-54a5-4915-9b0c-bc668fe625ff-kube-api-access-bcdbd\") pod \"olm-operator-6b444d44fb-xqsdv\" (UID: \"a2ade101-54a5-4915-9b0c-bc668fe625ff\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742619 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksrvf\" (UniqueName: \"kubernetes.io/projected/daadac3d-d689-4424-a6ff-1a6da646d03f-kube-api-access-ksrvf\") pod \"downloads-7954f5f757-glf8c\" (UID: \"daadac3d-d689-4424-a6ff-1a6da646d03f\") " pod="openshift-console/downloads-7954f5f757-glf8c" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742643 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2558ee2-4e5f-419a-94ca-5b37d181f83c-config\") pod \"kube-controller-manager-operator-78b949d7b-zcgr5\" (UID: \"f2558ee2-4e5f-419a-94ca-5b37d181f83c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742665 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-oauth-serving-cert\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742693 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-console-config\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742722 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/27255302-b4b6-40dc-a5fd-f6e376b7e05d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742765 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t6qkh\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.742788 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.743172 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/27255302-b4b6-40dc-a5fd-f6e376b7e05d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.744265 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.744293 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4cec2f60-1603-4a04-baab-3d9c070257da-serving-cert\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: E1215 12:19:38.745038 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.245025012 +0000 UTC m=+140.187318042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.750839 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-certificates\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.753175 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-tls\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.754823 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/27255302-b4b6-40dc-a5fd-f6e376b7e05d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.757071 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dqww\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-kube-api-access-6dqww\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.773432 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-bound-sa-token\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.790294 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-trusted-ca\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845191 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845347 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcdbd\" (UniqueName: \"kubernetes.io/projected/a2ade101-54a5-4915-9b0c-bc668fe625ff-kube-api-access-bcdbd\") pod \"olm-operator-6b444d44fb-xqsdv\" (UID: \"a2ade101-54a5-4915-9b0c-bc668fe625ff\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845376 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kfvp\" (UniqueName: \"kubernetes.io/projected/94ce3699-c8e3-42df-bafe-8cdd7572c406-kube-api-access-6kfvp\") pod \"migrator-59844c95c7-4cwjp\" (UID: \"94ce3699-c8e3-42df-bafe-8cdd7572c406\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845403 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksrvf\" (UniqueName: \"kubernetes.io/projected/daadac3d-d689-4424-a6ff-1a6da646d03f-kube-api-access-ksrvf\") pod \"downloads-7954f5f757-glf8c\" (UID: \"daadac3d-d689-4424-a6ff-1a6da646d03f\") " pod="openshift-console/downloads-7954f5f757-glf8c" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845421 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2558ee2-4e5f-419a-94ca-5b37d181f83c-config\") pod \"kube-controller-manager-operator-78b949d7b-zcgr5\" (UID: \"f2558ee2-4e5f-419a-94ca-5b37d181f83c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845439 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-oauth-serving-cert\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845453 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-console-config\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845497 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t6qkh\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845513 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845527 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4cec2f60-1603-4a04-baab-3d9c070257da-serving-cert\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845545 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845560 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fd3ac7a-862d-45ff-a931-5962b105ac6e-serving-cert\") pod \"service-ca-operator-777779d784-lsvbn\" (UID: \"4fd3ac7a-862d-45ff-a931-5962b105ac6e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845573 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1-metrics-tls\") pod \"dns-operator-744455d44c-9twm9\" (UID: \"e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1\") " pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845591 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bda713e9-91f5-40bc-806e-205f4a5dae25-trusted-ca\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845606 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4cec2f60-1603-4a04-baab-3d9c070257da-etcd-ca\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845620 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hblb8\" (UniqueName: \"kubernetes.io/projected/392f2574-b01c-41bf-b0df-7774696560ee-kube-api-access-hblb8\") pod \"dns-default-827sb\" (UID: \"392f2574-b01c-41bf-b0df-7774696560ee\") " pod="openshift-dns/dns-default-827sb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845636 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/392f2574-b01c-41bf-b0df-7774696560ee-config-volume\") pod \"dns-default-827sb\" (UID: \"392f2574-b01c-41bf-b0df-7774696560ee\") " pod="openshift-dns/dns-default-827sb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845650 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bda713e9-91f5-40bc-806e-205f4a5dae25-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845668 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/32919d63-a33b-44ac-af4e-aeed2e97a3ac-tmpfs\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845682 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc14ce9a-a4fe-4be7-9fdd-e6e75e415829-cert\") pod \"ingress-canary-vchb6\" (UID: \"fc14ce9a-a4fe-4be7-9fdd-e6e75e415829\") " pod="openshift-ingress-canary/ingress-canary-vchb6" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845698 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e-proxy-tls\") pod \"machine-config-controller-84d6567774-zhwtq\" (UID: \"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845713 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b7161e6-9af7-4443-9af7-2d11e0f36d68-secret-volume\") pod \"collect-profiles-29430015-f8vkb\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845730 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk5d2\" (UniqueName: \"kubernetes.io/projected/1699adb2-2c5e-4345-92bc-184205052fc8-kube-api-access-wk5d2\") pod \"kube-storage-version-migrator-operator-b67b599dd-ts787\" (UID: \"1699adb2-2c5e-4345-92bc-184205052fc8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845744 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b7161e6-9af7-4443-9af7-2d11e0f36d68-config-volume\") pod \"collect-profiles-29430015-f8vkb\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845780 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zhwtq\" (UID: \"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845805 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cec2f60-1603-4a04-baab-3d9c070257da-config\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845828 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn6pj\" (UniqueName: \"kubernetes.io/projected/bda713e9-91f5-40bc-806e-205f4a5dae25-kube-api-access-tn6pj\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845867 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v2sq\" (UniqueName: \"kubernetes.io/projected/aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e-kube-api-access-4v2sq\") pod \"machine-config-controller-84d6567774-zhwtq\" (UID: \"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845907 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1b3e2612-1a37-4784-940f-44cb98fa747a-node-bootstrap-token\") pod \"machine-config-server-ls926\" (UID: \"1b3e2612-1a37-4784-940f-44cb98fa747a\") " pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845925 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbszz\" (UniqueName: \"kubernetes.io/projected/4cec2f60-1603-4a04-baab-3d9c070257da-kube-api-access-dbszz\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845951 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2558ee2-4e5f-419a-94ca-5b37d181f83c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zcgr5\" (UID: \"f2558ee2-4e5f-419a-94ca-5b37d181f83c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845972 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-oauth-config\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.845986 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4cec2f60-1603-4a04-baab-3d9c070257da-etcd-service-ca\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846005 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-plugins-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846030 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-mountpoint-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846044 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d34646d2-f56c-4b6a-bbaa-c9f3485b61e0-signing-key\") pod \"service-ca-9c57cc56f-94r2g\" (UID: \"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0\") " pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846062 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb7kh\" (UniqueName: \"kubernetes.io/projected/556c5cc5-d57e-413d-aa79-421d196cb799-kube-api-access-fb7kh\") pod \"marketplace-operator-79b997595-t6qkh\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846077 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d561cf62-c184-4256-bec7-af48aa15afcb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mnl8k\" (UID: \"d561cf62-c184-4256-bec7-af48aa15afcb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846092 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd3ac7a-862d-45ff-a931-5962b105ac6e-config\") pod \"service-ca-operator-777779d784-lsvbn\" (UID: \"4fd3ac7a-862d-45ff-a931-5962b105ac6e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846120 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bda713e9-91f5-40bc-806e-205f4a5dae25-metrics-tls\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846135 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgsfx\" (UniqueName: \"kubernetes.io/projected/273927d9-dcea-4e2c-8c0c-c80cb6902cda-kube-api-access-lgsfx\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846161 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/994235ed-c7e5-44d3-bef7-c8df6b3224a3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846184 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9rhq\" (UniqueName: \"kubernetes.io/projected/790b2072-3dd0-4e6c-896c-724a3cd2bc5d-kube-api-access-r9rhq\") pod \"package-server-manager-789f6589d5-8wqj5\" (UID: \"790b2072-3dd0-4e6c-896c-724a3cd2bc5d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846199 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tfmj\" (UniqueName: \"kubernetes.io/projected/674241e5-eef3-4cf8-b4ff-c14a7a20d1b4-kube-api-access-2tfmj\") pod \"catalog-operator-68c6474976-txqnw\" (UID: \"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846216 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-serving-cert\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846232 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77kjq\" (UniqueName: \"kubernetes.io/projected/d34646d2-f56c-4b6a-bbaa-c9f3485b61e0-kube-api-access-77kjq\") pod \"service-ca-9c57cc56f-94r2g\" (UID: \"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0\") " pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846249 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-registration-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846266 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-trusted-ca-bundle\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846283 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgs4m\" (UniqueName: \"kubernetes.io/projected/32919d63-a33b-44ac-af4e-aeed2e97a3ac-kube-api-access-xgs4m\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846301 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1699adb2-2c5e-4345-92bc-184205052fc8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ts787\" (UID: \"1699adb2-2c5e-4345-92bc-184205052fc8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846319 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846333 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-csi-data-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846350 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57hz2\" (UniqueName: \"kubernetes.io/projected/fc14ce9a-a4fe-4be7-9fdd-e6e75e415829-kube-api-access-57hz2\") pod \"ingress-canary-vchb6\" (UID: \"fc14ce9a-a4fe-4be7-9fdd-e6e75e415829\") " pod="openshift-ingress-canary/ingress-canary-vchb6" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846368 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp24b\" (UniqueName: \"kubernetes.io/projected/e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1-kube-api-access-hp24b\") pod \"dns-operator-744455d44c-9twm9\" (UID: \"e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1\") " pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846397 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/994235ed-c7e5-44d3-bef7-c8df6b3224a3-proxy-tls\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846412 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a2ade101-54a5-4915-9b0c-bc668fe625ff-srv-cert\") pod \"olm-operator-6b444d44fb-xqsdv\" (UID: \"a2ade101-54a5-4915-9b0c-bc668fe625ff\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846433 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d620f9fd-1374-4f93-972d-93f2b9173471-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ngv2q\" (UID: \"d620f9fd-1374-4f93-972d-93f2b9173471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846450 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-socket-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846464 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32919d63-a33b-44ac-af4e-aeed2e97a3ac-apiservice-cert\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846489 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t6qkh\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846503 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/994235ed-c7e5-44d3-bef7-c8df6b3224a3-images\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846520 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/674241e5-eef3-4cf8-b4ff-c14a7a20d1b4-srv-cert\") pod \"catalog-operator-68c6474976-txqnw\" (UID: \"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846546 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w4mc\" (UniqueName: \"kubernetes.io/projected/4fd3ac7a-862d-45ff-a931-5962b105ac6e-kube-api-access-8w4mc\") pod \"service-ca-operator-777779d784-lsvbn\" (UID: \"4fd3ac7a-862d-45ff-a931-5962b105ac6e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846563 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a2ade101-54a5-4915-9b0c-bc668fe625ff-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xqsdv\" (UID: \"a2ade101-54a5-4915-9b0c-bc668fe625ff\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846584 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26xtv\" (UniqueName: \"kubernetes.io/projected/994235ed-c7e5-44d3-bef7-c8df6b3224a3-kube-api-access-26xtv\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846605 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrffv\" (UniqueName: \"kubernetes.io/projected/c8458395-1102-446b-8b75-752820a9cfdb-kube-api-access-vrffv\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846626 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-service-ca\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846644 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1699adb2-2c5e-4345-92bc-184205052fc8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ts787\" (UID: \"1699adb2-2c5e-4345-92bc-184205052fc8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846666 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v4tl\" (UniqueName: \"kubernetes.io/projected/d620f9fd-1374-4f93-972d-93f2b9173471-kube-api-access-2v4tl\") pod \"control-plane-machine-set-operator-78cbb6b69f-ngv2q\" (UID: \"d620f9fd-1374-4f93-972d-93f2b9173471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846689 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/790b2072-3dd0-4e6c-896c-724a3cd2bc5d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8wqj5\" (UID: \"790b2072-3dd0-4e6c-896c-724a3cd2bc5d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846710 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1b3e2612-1a37-4784-940f-44cb98fa747a-certs\") pod \"machine-config-server-ls926\" (UID: \"1b3e2612-1a37-4784-940f-44cb98fa747a\") " pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846736 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w88c6\" (UniqueName: \"kubernetes.io/projected/6b7161e6-9af7-4443-9af7-2d11e0f36d68-kube-api-access-w88c6\") pod \"collect-profiles-29430015-f8vkb\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846769 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d34646d2-f56c-4b6a-bbaa-c9f3485b61e0-signing-cabundle\") pod \"service-ca-9c57cc56f-94r2g\" (UID: \"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0\") " pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846790 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4cec2f60-1603-4a04-baab-3d9c070257da-etcd-client\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846811 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tw7l\" (UniqueName: \"kubernetes.io/projected/1b3e2612-1a37-4784-940f-44cb98fa747a-kube-api-access-8tw7l\") pod \"machine-config-server-ls926\" (UID: \"1b3e2612-1a37-4784-940f-44cb98fa747a\") " pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846847 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4rwv\" (UniqueName: \"kubernetes.io/projected/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-kube-api-access-d4rwv\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.846929 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2558ee2-4e5f-419a-94ca-5b37d181f83c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zcgr5\" (UID: \"f2558ee2-4e5f-419a-94ca-5b37d181f83c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:38 crc kubenswrapper[4719]: E1215 12:19:38.846951 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.346932117 +0000 UTC m=+140.289225147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.847836 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2558ee2-4e5f-419a-94ca-5b37d181f83c-config\") pod \"kube-controller-manager-operator-78b949d7b-zcgr5\" (UID: \"f2558ee2-4e5f-419a-94ca-5b37d181f83c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.848677 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-oauth-serving-cert\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.849161 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-console-config\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.849512 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32919d63-a33b-44ac-af4e-aeed2e97a3ac-webhook-cert\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.849558 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/392f2574-b01c-41bf-b0df-7774696560ee-metrics-tls\") pod \"dns-default-827sb\" (UID: \"392f2574-b01c-41bf-b0df-7774696560ee\") " pod="openshift-dns/dns-default-827sb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.849599 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlmjp\" (UniqueName: \"kubernetes.io/projected/d561cf62-c184-4256-bec7-af48aa15afcb-kube-api-access-rlmjp\") pod \"multus-admission-controller-857f4d67dd-mnl8k\" (UID: \"d561cf62-c184-4256-bec7-af48aa15afcb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.849616 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/674241e5-eef3-4cf8-b4ff-c14a7a20d1b4-profile-collector-cert\") pod \"catalog-operator-68c6474976-txqnw\" (UID: \"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.850374 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t6qkh\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.853243 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fd3ac7a-862d-45ff-a931-5962b105ac6e-config\") pod \"service-ca-operator-777779d784-lsvbn\" (UID: \"4fd3ac7a-862d-45ff-a931-5962b105ac6e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.853703 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/994235ed-c7e5-44d3-bef7-c8df6b3224a3-images\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.856006 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-service-ca\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.863519 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zhwtq\" (UID: \"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.864244 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cec2f60-1603-4a04-baab-3d9c070257da-config\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.870697 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t6qkh\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.871051 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.872117 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4cec2f60-1603-4a04-baab-3d9c070257da-etcd-ca\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.873574 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/994235ed-c7e5-44d3-bef7-c8df6b3224a3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.891992 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1-metrics-tls\") pod \"dns-operator-744455d44c-9twm9\" (UID: \"e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1\") " pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.904199 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/790b2072-3dd0-4e6c-896c-724a3cd2bc5d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8wqj5\" (UID: \"790b2072-3dd0-4e6c-896c-724a3cd2bc5d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.904515 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4cec2f60-1603-4a04-baab-3d9c070257da-etcd-service-ca\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.905325 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-trusted-ca-bundle\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.906079 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a2ade101-54a5-4915-9b0c-bc668fe625ff-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xqsdv\" (UID: \"a2ade101-54a5-4915-9b0c-bc668fe625ff\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.907755 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e-proxy-tls\") pod \"machine-config-controller-84d6567774-zhwtq\" (UID: \"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.915685 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb7kh\" (UniqueName: \"kubernetes.io/projected/556c5cc5-d57e-413d-aa79-421d196cb799-kube-api-access-fb7kh\") pod \"marketplace-operator-79b997595-t6qkh\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.916710 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2558ee2-4e5f-419a-94ca-5b37d181f83c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zcgr5\" (UID: \"f2558ee2-4e5f-419a-94ca-5b37d181f83c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.920245 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrffv\" (UniqueName: \"kubernetes.io/projected/c8458395-1102-446b-8b75-752820a9cfdb-kube-api-access-vrffv\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.925446 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4cec2f60-1603-4a04-baab-3d9c070257da-serving-cert\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.926105 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-serving-cert\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.927543 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-oauth-config\") pod \"console-f9d7485db-2xxrl\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.928018 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v4tl\" (UniqueName: \"kubernetes.io/projected/d620f9fd-1374-4f93-972d-93f2b9173471-kube-api-access-2v4tl\") pod \"control-plane-machine-set-operator-78cbb6b69f-ngv2q\" (UID: \"d620f9fd-1374-4f93-972d-93f2b9173471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.929192 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcdbd\" (UniqueName: \"kubernetes.io/projected/a2ade101-54a5-4915-9b0c-bc668fe625ff-kube-api-access-bcdbd\") pod \"olm-operator-6b444d44fb-xqsdv\" (UID: \"a2ade101-54a5-4915-9b0c-bc668fe625ff\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.931497 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksrvf\" (UniqueName: \"kubernetes.io/projected/daadac3d-d689-4424-a6ff-1a6da646d03f-kube-api-access-ksrvf\") pod \"downloads-7954f5f757-glf8c\" (UID: \"daadac3d-d689-4424-a6ff-1a6da646d03f\") " pod="openshift-console/downloads-7954f5f757-glf8c" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.940486 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.940617 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a2ade101-54a5-4915-9b0c-bc668fe625ff-srv-cert\") pod \"olm-operator-6b444d44fb-xqsdv\" (UID: \"a2ade101-54a5-4915-9b0c-bc668fe625ff\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.940828 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4cec2f60-1603-4a04-baab-3d9c070257da-etcd-client\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.942315 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4fd3ac7a-862d-45ff-a931-5962b105ac6e-serving-cert\") pod \"service-ca-operator-777779d784-lsvbn\" (UID: \"4fd3ac7a-862d-45ff-a931-5962b105ac6e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.946434 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/994235ed-c7e5-44d3-bef7-c8df6b3224a3-proxy-tls\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.946508 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w4mc\" (UniqueName: \"kubernetes.io/projected/4fd3ac7a-862d-45ff-a931-5962b105ac6e-kube-api-access-8w4mc\") pod \"service-ca-operator-777779d784-lsvbn\" (UID: \"4fd3ac7a-862d-45ff-a931-5962b105ac6e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.948439 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v2sq\" (UniqueName: \"kubernetes.io/projected/aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e-kube-api-access-4v2sq\") pod \"machine-config-controller-84d6567774-zhwtq\" (UID: \"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.948510 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.948796 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f2558ee2-4e5f-419a-94ca-5b37d181f83c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zcgr5\" (UID: \"f2558ee2-4e5f-419a-94ca-5b37d181f83c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.949826 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp24b\" (UniqueName: \"kubernetes.io/projected/e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1-kube-api-access-hp24b\") pod \"dns-operator-744455d44c-9twm9\" (UID: \"e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1\") " pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950472 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk5d2\" (UniqueName: \"kubernetes.io/projected/1699adb2-2c5e-4345-92bc-184205052fc8-kube-api-access-wk5d2\") pod \"kube-storage-version-migrator-operator-b67b599dd-ts787\" (UID: \"1699adb2-2c5e-4345-92bc-184205052fc8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950499 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b7161e6-9af7-4443-9af7-2d11e0f36d68-config-volume\") pod \"collect-profiles-29430015-f8vkb\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950522 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn6pj\" (UniqueName: \"kubernetes.io/projected/bda713e9-91f5-40bc-806e-205f4a5dae25-kube-api-access-tn6pj\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950548 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1b3e2612-1a37-4784-940f-44cb98fa747a-node-bootstrap-token\") pod \"machine-config-server-ls926\" (UID: \"1b3e2612-1a37-4784-940f-44cb98fa747a\") " pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950579 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-plugins-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950597 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-mountpoint-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950615 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d34646d2-f56c-4b6a-bbaa-c9f3485b61e0-signing-key\") pod \"service-ca-9c57cc56f-94r2g\" (UID: \"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0\") " pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950630 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bda713e9-91f5-40bc-806e-205f4a5dae25-metrics-tls\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950648 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d561cf62-c184-4256-bec7-af48aa15afcb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mnl8k\" (UID: \"d561cf62-c184-4256-bec7-af48aa15afcb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950662 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgsfx\" (UniqueName: \"kubernetes.io/projected/273927d9-dcea-4e2c-8c0c-c80cb6902cda-kube-api-access-lgsfx\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950684 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tfmj\" (UniqueName: \"kubernetes.io/projected/674241e5-eef3-4cf8-b4ff-c14a7a20d1b4-kube-api-access-2tfmj\") pod \"catalog-operator-68c6474976-txqnw\" (UID: \"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950702 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77kjq\" (UniqueName: \"kubernetes.io/projected/d34646d2-f56c-4b6a-bbaa-c9f3485b61e0-kube-api-access-77kjq\") pod \"service-ca-9c57cc56f-94r2g\" (UID: \"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0\") " pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950718 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-registration-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950735 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgs4m\" (UniqueName: \"kubernetes.io/projected/32919d63-a33b-44ac-af4e-aeed2e97a3ac-kube-api-access-xgs4m\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950753 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1699adb2-2c5e-4345-92bc-184205052fc8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ts787\" (UID: \"1699adb2-2c5e-4345-92bc-184205052fc8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950770 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-csi-data-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950786 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57hz2\" (UniqueName: \"kubernetes.io/projected/fc14ce9a-a4fe-4be7-9fdd-e6e75e415829-kube-api-access-57hz2\") pod \"ingress-canary-vchb6\" (UID: \"fc14ce9a-a4fe-4be7-9fdd-e6e75e415829\") " pod="openshift-ingress-canary/ingress-canary-vchb6" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950809 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950824 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-socket-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950838 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32919d63-a33b-44ac-af4e-aeed2e97a3ac-apiservice-cert\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950900 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/674241e5-eef3-4cf8-b4ff-c14a7a20d1b4-srv-cert\") pod \"catalog-operator-68c6474976-txqnw\" (UID: \"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950933 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1699adb2-2c5e-4345-92bc-184205052fc8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ts787\" (UID: \"1699adb2-2c5e-4345-92bc-184205052fc8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950948 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1b3e2612-1a37-4784-940f-44cb98fa747a-certs\") pod \"machine-config-server-ls926\" (UID: \"1b3e2612-1a37-4784-940f-44cb98fa747a\") " pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950963 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w88c6\" (UniqueName: \"kubernetes.io/projected/6b7161e6-9af7-4443-9af7-2d11e0f36d68-kube-api-access-w88c6\") pod \"collect-profiles-29430015-f8vkb\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950978 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d34646d2-f56c-4b6a-bbaa-c9f3485b61e0-signing-cabundle\") pod \"service-ca-9c57cc56f-94r2g\" (UID: \"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0\") " pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.950993 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tw7l\" (UniqueName: \"kubernetes.io/projected/1b3e2612-1a37-4784-940f-44cb98fa747a-kube-api-access-8tw7l\") pod \"machine-config-server-ls926\" (UID: \"1b3e2612-1a37-4784-940f-44cb98fa747a\") " pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951008 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32919d63-a33b-44ac-af4e-aeed2e97a3ac-webhook-cert\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951030 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/392f2574-b01c-41bf-b0df-7774696560ee-metrics-tls\") pod \"dns-default-827sb\" (UID: \"392f2574-b01c-41bf-b0df-7774696560ee\") " pod="openshift-dns/dns-default-827sb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951046 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlmjp\" (UniqueName: \"kubernetes.io/projected/d561cf62-c184-4256-bec7-af48aa15afcb-kube-api-access-rlmjp\") pod \"multus-admission-controller-857f4d67dd-mnl8k\" (UID: \"d561cf62-c184-4256-bec7-af48aa15afcb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951060 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/674241e5-eef3-4cf8-b4ff-c14a7a20d1b4-profile-collector-cert\") pod \"catalog-operator-68c6474976-txqnw\" (UID: \"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951077 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kfvp\" (UniqueName: \"kubernetes.io/projected/94ce3699-c8e3-42df-bafe-8cdd7572c406-kube-api-access-6kfvp\") pod \"migrator-59844c95c7-4cwjp\" (UID: \"94ce3699-c8e3-42df-bafe-8cdd7572c406\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951120 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hblb8\" (UniqueName: \"kubernetes.io/projected/392f2574-b01c-41bf-b0df-7774696560ee-kube-api-access-hblb8\") pod \"dns-default-827sb\" (UID: \"392f2574-b01c-41bf-b0df-7774696560ee\") " pod="openshift-dns/dns-default-827sb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951140 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bda713e9-91f5-40bc-806e-205f4a5dae25-trusted-ca\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951158 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/392f2574-b01c-41bf-b0df-7774696560ee-config-volume\") pod \"dns-default-827sb\" (UID: \"392f2574-b01c-41bf-b0df-7774696560ee\") " pod="openshift-dns/dns-default-827sb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951191 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bda713e9-91f5-40bc-806e-205f4a5dae25-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951238 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/32919d63-a33b-44ac-af4e-aeed2e97a3ac-tmpfs\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951260 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b7161e6-9af7-4443-9af7-2d11e0f36d68-secret-volume\") pod \"collect-profiles-29430015-f8vkb\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951273 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc14ce9a-a4fe-4be7-9fdd-e6e75e415829-cert\") pod \"ingress-canary-vchb6\" (UID: \"fc14ce9a-a4fe-4be7-9fdd-e6e75e415829\") " pod="openshift-ingress-canary/ingress-canary-vchb6" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.951798 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-socket-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.953301 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d620f9fd-1374-4f93-972d-93f2b9173471-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ngv2q\" (UID: \"d620f9fd-1374-4f93-972d-93f2b9173471\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.955416 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b7161e6-9af7-4443-9af7-2d11e0f36d68-config-volume\") pod \"collect-profiles-29430015-f8vkb\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.956879 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-registration-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.957064 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1699adb2-2c5e-4345-92bc-184205052fc8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ts787\" (UID: \"1699adb2-2c5e-4345-92bc-184205052fc8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.957150 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-csi-data-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: E1215 12:19:38.957485 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.457473619 +0000 UTC m=+140.399766649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.961374 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-plugins-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.961470 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/273927d9-dcea-4e2c-8c0c-c80cb6902cda-mountpoint-dir\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.966100 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bda713e9-91f5-40bc-806e-205f4a5dae25-trusted-ca\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.966372 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/32919d63-a33b-44ac-af4e-aeed2e97a3ac-tmpfs\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.967948 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26xtv\" (UniqueName: \"kubernetes.io/projected/994235ed-c7e5-44d3-bef7-c8df6b3224a3-kube-api-access-26xtv\") pod \"machine-config-operator-74547568cd-z56dp\" (UID: \"994235ed-c7e5-44d3-bef7-c8df6b3224a3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.968677 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9rhq\" (UniqueName: \"kubernetes.io/projected/790b2072-3dd0-4e6c-896c-724a3cd2bc5d-kube-api-access-r9rhq\") pod \"package-server-manager-789f6589d5-8wqj5\" (UID: \"790b2072-3dd0-4e6c-896c-724a3cd2bc5d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.969454 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d34646d2-f56c-4b6a-bbaa-c9f3485b61e0-signing-cabundle\") pod \"service-ca-9c57cc56f-94r2g\" (UID: \"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0\") " pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.970515 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/392f2574-b01c-41bf-b0df-7774696560ee-config-volume\") pod \"dns-default-827sb\" (UID: \"392f2574-b01c-41bf-b0df-7774696560ee\") " pod="openshift-dns/dns-default-827sb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.972301 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbszz\" (UniqueName: \"kubernetes.io/projected/4cec2f60-1603-4a04-baab-3d9c070257da-kube-api-access-dbszz\") pod \"etcd-operator-b45778765-7g9lh\" (UID: \"4cec2f60-1603-4a04-baab-3d9c070257da\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.976496 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b7161e6-9af7-4443-9af7-2d11e0f36d68-secret-volume\") pod \"collect-profiles-29430015-f8vkb\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.976985 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc14ce9a-a4fe-4be7-9fdd-e6e75e415829-cert\") pod \"ingress-canary-vchb6\" (UID: \"fc14ce9a-a4fe-4be7-9fdd-e6e75e415829\") " pod="openshift-ingress-canary/ingress-canary-vchb6" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.980076 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f"] Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.982526 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1b3e2612-1a37-4784-940f-44cb98fa747a-node-bootstrap-token\") pod \"machine-config-server-ls926\" (UID: \"1b3e2612-1a37-4784-940f-44cb98fa747a\") " pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.983720 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/392f2574-b01c-41bf-b0df-7774696560ee-metrics-tls\") pod \"dns-default-827sb\" (UID: \"392f2574-b01c-41bf-b0df-7774696560ee\") " pod="openshift-dns/dns-default-827sb" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.983887 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d34646d2-f56c-4b6a-bbaa-c9f3485b61e0-signing-key\") pod \"service-ca-9c57cc56f-94r2g\" (UID: \"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0\") " pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.983889 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/674241e5-eef3-4cf8-b4ff-c14a7a20d1b4-profile-collector-cert\") pod \"catalog-operator-68c6474976-txqnw\" (UID: \"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.985274 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk5d2\" (UniqueName: \"kubernetes.io/projected/1699adb2-2c5e-4345-92bc-184205052fc8-kube-api-access-wk5d2\") pod \"kube-storage-version-migrator-operator-b67b599dd-ts787\" (UID: \"1699adb2-2c5e-4345-92bc-184205052fc8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.985702 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4rwv\" (UniqueName: \"kubernetes.io/projected/8a457b5d-1084-4330-bcf5-bd5c5ef97c85-kube-api-access-d4rwv\") pod \"cluster-image-registry-operator-dc59b4c8b-2l8f2\" (UID: \"8a457b5d-1084-4330-bcf5-bd5c5ef97c85\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.986441 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/674241e5-eef3-4cf8-b4ff-c14a7a20d1b4-srv-cert\") pod \"catalog-operator-68c6474976-txqnw\" (UID: \"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.986964 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bda713e9-91f5-40bc-806e-205f4a5dae25-metrics-tls\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.993131 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32919d63-a33b-44ac-af4e-aeed2e97a3ac-apiservice-cert\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.994644 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgs4m\" (UniqueName: \"kubernetes.io/projected/32919d63-a33b-44ac-af4e-aeed2e97a3ac-kube-api-access-xgs4m\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.995675 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1699adb2-2c5e-4345-92bc-184205052fc8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ts787\" (UID: \"1699adb2-2c5e-4345-92bc-184205052fc8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:38 crc kubenswrapper[4719]: I1215 12:19:38.996193 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d561cf62-c184-4256-bec7-af48aa15afcb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mnl8k\" (UID: \"d561cf62-c184-4256-bec7-af48aa15afcb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:38.997366 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1b3e2612-1a37-4784-940f-44cb98fa747a-certs\") pod \"machine-config-server-ls926\" (UID: \"1b3e2612-1a37-4784-940f-44cb98fa747a\") " pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:38.997643 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:38.997914 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32919d63-a33b-44ac-af4e-aeed2e97a3ac-webhook-cert\") pod \"packageserver-d55dfcdfc-5gcx2\" (UID: \"32919d63-a33b-44ac-af4e-aeed2e97a3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.005021 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgsfx\" (UniqueName: \"kubernetes.io/projected/273927d9-dcea-4e2c-8c0c-c80cb6902cda-kube-api-access-lgsfx\") pod \"csi-hostpathplugin-4mbgz\" (UID: \"273927d9-dcea-4e2c-8c0c-c80cb6902cda\") " pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.033911 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.034627 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn6pj\" (UniqueName: \"kubernetes.io/projected/bda713e9-91f5-40bc-806e-205f4a5dae25-kube-api-access-tn6pj\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.051891 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.052326 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.552297979 +0000 UTC m=+140.494591009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.052493 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.053070 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.553055084 +0000 UTC m=+140.495348114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.057987 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-glf8c" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.065911 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.066907 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tfmj\" (UniqueName: \"kubernetes.io/projected/674241e5-eef3-4cf8-b4ff-c14a7a20d1b4-kube-api-access-2tfmj\") pod \"catalog-operator-68c6474976-txqnw\" (UID: \"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.074896 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.083143 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77kjq\" (UniqueName: \"kubernetes.io/projected/d34646d2-f56c-4b6a-bbaa-c9f3485b61e0-kube-api-access-77kjq\") pod \"service-ca-9c57cc56f-94r2g\" (UID: \"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0\") " pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.083917 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.089265 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.098644 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57hz2\" (UniqueName: \"kubernetes.io/projected/fc14ce9a-a4fe-4be7-9fdd-e6e75e415829-kube-api-access-57hz2\") pod \"ingress-canary-vchb6\" (UID: \"fc14ce9a-a4fe-4be7-9fdd-e6e75e415829\") " pod="openshift-ingress-canary/ingress-canary-vchb6" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.098847 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.108147 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.126769 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.134640 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kfvp\" (UniqueName: \"kubernetes.io/projected/94ce3699-c8e3-42df-bafe-8cdd7572c406-kube-api-access-6kfvp\") pod \"migrator-59844c95c7-4cwjp\" (UID: \"94ce3699-c8e3-42df-bafe-8cdd7572c406\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.157342 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.159354 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.159500 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.177280 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hblb8\" (UniqueName: \"kubernetes.io/projected/392f2574-b01c-41bf-b0df-7774696560ee-kube-api-access-hblb8\") pod \"dns-default-827sb\" (UID: \"392f2574-b01c-41bf-b0df-7774696560ee\") " pod="openshift-dns/dns-default-827sb" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.159845 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.160508 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.160586 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.660566235 +0000 UTC m=+140.602859265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.164693 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.172537 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bda713e9-91f5-40bc-806e-205f4a5dae25-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7z22d\" (UID: \"bda713e9-91f5-40bc-806e-205f4a5dae25\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.159720 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.181703 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.181805 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-vchb6" Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.182250 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.682237135 +0000 UTC m=+140.624530165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.188505 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w88c6\" (UniqueName: \"kubernetes.io/projected/6b7161e6-9af7-4443-9af7-2d11e0f36d68-kube-api-access-w88c6\") pod \"collect-profiles-29430015-f8vkb\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.188834 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-827sb" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.200353 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.206551 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tw7l\" (UniqueName: \"kubernetes.io/projected/1b3e2612-1a37-4784-940f-44cb98fa747a-kube-api-access-8tw7l\") pod \"machine-config-server-ls926\" (UID: \"1b3e2612-1a37-4784-940f-44cb98fa747a\") " pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.214344 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-7lswj"] Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.214674 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.224933 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.235739 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlmjp\" (UniqueName: \"kubernetes.io/projected/d561cf62-c184-4256-bec7-af48aa15afcb-kube-api-access-rlmjp\") pod \"multus-admission-controller-857f4d67dd-mnl8k\" (UID: \"d561cf62-c184-4256-bec7-af48aa15afcb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.241031 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r"] Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.251212 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkz2j"] Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.275247 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp"] Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.282383 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.282549 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.782525546 +0000 UTC m=+140.724818576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.384493 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.385082 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.885071902 +0000 UTC m=+140.827364932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.424185 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2"] Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.429912 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.436734 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.466474 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5"] Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.474177 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.485395 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.485825 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:39.985809038 +0000 UTC m=+140.928102068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.495126 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm"] Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.495270 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ls926" Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.517304 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p"] Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.518669 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs"] Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.532716 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486"] Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.586727 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.587051 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.087037711 +0000 UTC m=+141.029330731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.636070 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" event={"ID":"0dd6e306-1355-43dd-895f-598d3d873a42","Type":"ContainerStarted","Data":"3afe2a0505d9cba605b3d790260c5433b42b3feb2d1c73e14da05db8045b96f9"} Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.638116 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" event={"ID":"30f62441-ac5d-48b7-b77c-8b6ff58546d1","Type":"ContainerStarted","Data":"c10425ec5fad031a2610c47091d55f1597ce40584440909fc6a291704717ffb1"} Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.639312 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" event={"ID":"5fe9cf63-1287-418f-b2c5-4e4e7458cf44","Type":"ContainerStarted","Data":"aae65cef2246ccafec9cf563c241796d847fae667522399f3442c503e8aeef4c"} Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.640081 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jkk5f" event={"ID":"99aced59-b6b7-4abc-8d73-2dc483a14f8f","Type":"ContainerStarted","Data":"9cc9338647e2e97035fef03dc42975c4b512b1c8fbb870cb64b682837cf6f4aa"} Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.640781 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" event={"ID":"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72","Type":"ContainerStarted","Data":"21e769618b16a31d4fbf3c92c1fbba8defe90f81c956ac4f386ab432e34f3782"} Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.641593 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" event={"ID":"e6665e73-ac68-4fcc-a0e0-85815184a0ea","Type":"ContainerStarted","Data":"65a7c1c640aced6bbb321c169e97f419b80ceaa9fe61ae6c575ca7f7e61bed92"} Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.642288 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" event={"ID":"bc81baea-ae9e-4c9d-af05-d15fd4b1591c","Type":"ContainerStarted","Data":"1cfd5e01f3eea3c5c034014111aafddab17f2c391a8aae0be84d2a2da72e4747"} Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.643110 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" event={"ID":"78ebc618-1e6f-47e9-9afe-b25bf76a19ff","Type":"ContainerStarted","Data":"ff6d75170f3bada13e6dd9d6000ccfb5db3757e1636c8755c678b5803897af30"} Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.687966 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.688304 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.188285894 +0000 UTC m=+141.130578924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: W1215 12:19:39.699058 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a457b5d_1084_4330_bcf5_bd5c5ef97c85.slice/crio-0d7cfe81930a4bf9acac018b1eacfbcf6e1453cb248fe27648a18e0815a03ab8 WatchSource:0}: Error finding container 0d7cfe81930a4bf9acac018b1eacfbcf6e1453cb248fe27648a18e0815a03ab8: Status 404 returned error can't find the container with id 0d7cfe81930a4bf9acac018b1eacfbcf6e1453cb248fe27648a18e0815a03ab8 Dec 15 12:19:39 crc kubenswrapper[4719]: W1215 12:19:39.703752 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod129630ef_67df_4003_aead_ea250b93c8b5.slice/crio-21576149ae6422ee85de3d99203f8a1a82c60275874f0e5bac81e799882ccfc5 WatchSource:0}: Error finding container 21576149ae6422ee85de3d99203f8a1a82c60275874f0e5bac81e799882ccfc5: Status 404 returned error can't find the container with id 21576149ae6422ee85de3d99203f8a1a82c60275874f0e5bac81e799882ccfc5 Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.790390 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.790729 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.290717586 +0000 UTC m=+141.233010606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.892769 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.892924 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.392907001 +0000 UTC m=+141.335200031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:39 crc kubenswrapper[4719]: I1215 12:19:39.893247 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:39 crc kubenswrapper[4719]: E1215 12:19:39.893617 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.393606664 +0000 UTC m=+141.335899694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:39.993983 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:39.994213 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.494198706 +0000 UTC m=+141.436491736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.044650 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t6qkh"] Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.095236 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.095598 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.595587633 +0000 UTC m=+141.537880663 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.198201 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.198512 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.698487032 +0000 UTC m=+141.640780072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.198650 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.198965 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.698952436 +0000 UTC m=+141.641245466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.205197 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q"] Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.258327 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-glf8c"] Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.271710 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp"] Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.285001 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5"] Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.299976 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.300317 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.800302263 +0000 UTC m=+141.742595293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.364630 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq"] Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.381091 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv"] Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.412968 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.413540 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:40.913526524 +0000 UTC m=+141.855819554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.514764 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.514994 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.014967383 +0000 UTC m=+141.957260413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.515413 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.515743 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.015730419 +0000 UTC m=+141.958023449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.617550 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.618075 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.118054918 +0000 UTC m=+142.060347958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.712136 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" event={"ID":"790b2072-3dd0-4e6c-896c-724a3cd2bc5d","Type":"ContainerStarted","Data":"15aa00193df944304f15662ed675682149f7a7413ce85dc2387143525e4dc987"} Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.735198 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.735477 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.235462858 +0000 UTC m=+142.177755888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.836069 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.836681 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.33666582 +0000 UTC m=+142.278958850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.845259 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" event={"ID":"bc81baea-ae9e-4c9d-af05-d15fd4b1591c","Type":"ContainerStarted","Data":"9a7c0362b3dcd50d3d457cd08fc3cdcfff709001210b076b9a8eddb5462a0f05"} Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.861928 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" event={"ID":"994235ed-c7e5-44d3-bef7-c8df6b3224a3","Type":"ContainerStarted","Data":"b939eb4ab724f4511aa7503efb2e224289008957d2b17d46d611efc4fa0c1f80"} Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.867057 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-glf8c" event={"ID":"daadac3d-d689-4424-a6ff-1a6da646d03f","Type":"ContainerStarted","Data":"a024af9ba9078525e48cd10b01da1a4c59cfd6241ac0d6bee85863dd170204b1"} Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.899818 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" event={"ID":"3a4fc82e-33f4-4b77-ad47-54554d0c08b4","Type":"ContainerStarted","Data":"9789ec7a3351896ee702265548fb7fefca78b09d57f98a25909160d267504f1c"} Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.940888 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:40 crc kubenswrapper[4719]: E1215 12:19:40.942516 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.442182344 +0000 UTC m=+142.384475374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.964008 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" event={"ID":"4d7d0aab-ec25-437b-99e5-4a39a14bac0c","Type":"ContainerStarted","Data":"42c0a5e2622a19d2a75f2e5ed3bdd9906c6f36e9a2fb770b76a34bc56ca69ed1"} Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.982328 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" event={"ID":"d620f9fd-1374-4f93-972d-93f2b9173471","Type":"ContainerStarted","Data":"ae6358b18284aa23cb5006793de4d7137a5f8c62427c185cb1b07caa5749d371"} Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.993928 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-29pwn" podStartSLOduration=117.993906242 podStartE2EDuration="1m57.993906242s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:40.876353017 +0000 UTC m=+141.818646047" watchObservedRunningTime="2025-12-15 12:19:40.993906242 +0000 UTC m=+141.936199292" Dec 15 12:19:40 crc kubenswrapper[4719]: I1215 12:19:40.995523 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-827sb"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.002653 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-94r2g"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.009835 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" event={"ID":"f2558ee2-4e5f-419a-94ca-5b37d181f83c","Type":"ContainerStarted","Data":"b08459e5f8c50616f9d020fcac0f3dff3a6c079c06869a2efc843a6e6e2aabad"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.034330 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" event={"ID":"0dd6e306-1355-43dd-895f-598d3d873a42","Type":"ContainerStarted","Data":"652885d66f6a99e9ff4cf2e87f1ab93e50d325f2a034efd049b95bc8c0c610c4"} Dec 15 12:19:41 crc kubenswrapper[4719]: E1215 12:19:41.042361 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.542331741 +0000 UTC m=+142.484624771 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.042539 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.042817 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:41 crc kubenswrapper[4719]: E1215 12:19:41.043251 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.543237081 +0000 UTC m=+142.485530121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.064548 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jkk5f" event={"ID":"99aced59-b6b7-4abc-8d73-2dc483a14f8f","Type":"ContainerStarted","Data":"774aab139df6ffa98e4f364c9493e5823f9860b0fc75cbe3e706983c8b3b88a9"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.073976 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" event={"ID":"b03b5a84-a26a-4564-9081-3cfa516c02b3","Type":"ContainerStarted","Data":"c8f80a232a13fb36e9b7aadabc0ccdc0b0b51339e5b30b81ba1e8abfb9161029"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.090249 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4mbgz"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.090500 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-jkk5f" podStartSLOduration=118.090490881 podStartE2EDuration="1m58.090490881s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:41.085766294 +0000 UTC m=+142.028059324" watchObservedRunningTime="2025-12-15 12:19:41.090490881 +0000 UTC m=+142.032783911" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.104171 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" event={"ID":"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51","Type":"ContainerStarted","Data":"27b1509b041d1fba741ed2ce4e953aba4d79b33af8ae9fcce3fefcc390056477"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.119577 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7g9lh"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.136205 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.147171 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:41 crc kubenswrapper[4719]: E1215 12:19:41.154457 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.654437855 +0000 UTC m=+142.596730885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.186328 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" event={"ID":"78ebc618-1e6f-47e9-9afe-b25bf76a19ff","Type":"ContainerStarted","Data":"d4da3e9853f3087159e921154d5155df96237f3f07b58eae3b51019e4be8bde5"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.211768 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fszmm" podStartSLOduration=118.211749929 podStartE2EDuration="1m58.211749929s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:41.209609647 +0000 UTC m=+142.151902677" watchObservedRunningTime="2025-12-15 12:19:41.211749929 +0000 UTC m=+142.154042949" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.215032 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9twm9"] Dec 15 12:19:41 crc kubenswrapper[4719]: W1215 12:19:41.233494 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod273927d9_dcea_4e2c_8c0c_c80cb6902cda.slice/crio-f0e391be48964945c561372c0bf554effa9ff02a0ad4d8b5dc79efd73b9fdc63 WatchSource:0}: Error finding container f0e391be48964945c561372c0bf554effa9ff02a0ad4d8b5dc79efd73b9fdc63: Status 404 returned error can't find the container with id f0e391be48964945c561372c0bf554effa9ff02a0ad4d8b5dc79efd73b9fdc63 Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.248435 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" event={"ID":"e6665e73-ac68-4fcc-a0e0-85815184a0ea","Type":"ContainerStarted","Data":"a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.254754 4719 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-nz9wh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.255169 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" podUID="e6665e73-ac68-4fcc-a0e0-85815184a0ea" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.258180 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.266554 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:41 crc kubenswrapper[4719]: E1215 12:19:41.267401 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.767379616 +0000 UTC m=+142.709672636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:41 crc kubenswrapper[4719]: W1215 12:19:41.273091 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fd3ac7a_862d_45ff_a931_5962b105ac6e.slice/crio-c5cd1c68095b528067692fee7f2f0d40e1b586a1d4e3f149db104398a36a7c0b WatchSource:0}: Error finding container c5cd1c68095b528067692fee7f2f0d40e1b586a1d4e3f149db104398a36a7c0b: Status 404 returned error can't find the container with id c5cd1c68095b528067692fee7f2f0d40e1b586a1d4e3f149db104398a36a7c0b Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.273868 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.283696 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-vchb6"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.297353 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.299048 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" podStartSLOduration=118.299031147 podStartE2EDuration="1m58.299031147s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:41.289262603 +0000 UTC m=+142.231555633" watchObservedRunningTime="2025-12-15 12:19:41.299031147 +0000 UTC m=+142.241324177" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.325590 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2xxrl"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.333134 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" event={"ID":"8f2d8d73-a1cc-4a67-9006-205567f3b907","Type":"ContainerStarted","Data":"d5ff7de1e950805363b0e764dcca2991e3b40e32135acc1cd926ed162fa6d534"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.374244 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:41 crc kubenswrapper[4719]: E1215 12:19:41.375307 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.875293141 +0000 UTC m=+142.817586171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.392379 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" event={"ID":"129630ef-67df-4003-aead-ea250b93c8b5","Type":"ContainerStarted","Data":"21576149ae6422ee85de3d99203f8a1a82c60275874f0e5bac81e799882ccfc5"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.403834 4719 generic.go:334] "Generic (PLEG): container finished" podID="30f62441-ac5d-48b7-b77c-8b6ff58546d1" containerID="86b9db90d7c0a5be3736a2668ca16a43b875909b5f63a19bc06e35e49dd2a6fe" exitCode=0 Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.403922 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" event={"ID":"30f62441-ac5d-48b7-b77c-8b6ff58546d1","Type":"ContainerDied","Data":"86b9db90d7c0a5be3736a2668ca16a43b875909b5f63a19bc06e35e49dd2a6fe"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.440680 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" podStartSLOduration=118.422164887 podStartE2EDuration="1m58.422164887s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:41.418252438 +0000 UTC m=+142.360545468" watchObservedRunningTime="2025-12-15 12:19:41.422164887 +0000 UTC m=+142.364457917" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.463081 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" event={"ID":"5fe9cf63-1287-418f-b2c5-4e4e7458cf44","Type":"ContainerStarted","Data":"f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.464085 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.477055 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:41 crc kubenswrapper[4719]: E1215 12:19:41.479160 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:41.97913113 +0000 UTC m=+142.921424160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.517137 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" podStartSLOduration=118.517119122 podStartE2EDuration="1m58.517119122s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:41.515593111 +0000 UTC m=+142.457886151" watchObservedRunningTime="2025-12-15 12:19:41.517119122 +0000 UTC m=+142.459412162" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.579421 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:41 crc kubenswrapper[4719]: E1215 12:19:41.580571 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:42.080554669 +0000 UTC m=+143.022847699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.588195 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7lswj" event={"ID":"3a91fe95-e935-4e7c-bee1-d3d291206f67","Type":"ContainerStarted","Data":"673d15393e646d2f0854893d0a7e18ae4fb7740e56479f105c342683d4c94129"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.588232 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.588245 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.588255 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.588510 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.591008 4719 patch_prober.go:28] interesting pod/console-operator-58897d9998-7lswj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/readyz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.591346 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7lswj" podUID="3a91fe95-e935-4e7c-bee1-d3d291206f67" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/readyz\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.594314 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" event={"ID":"b44673b6-1b68-46e8-9b5f-b14c3389a017","Type":"ContainerStarted","Data":"34cfc63ae91f0ab43b21cc5835e257789493865bd3926acf76a7657fd7eba62a"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.636672 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:41 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:41 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:41 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.636759 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.622513 4719 generic.go:334] "Generic (PLEG): container finished" podID="8b58d079-d1d0-4923-8ba3-bfeb1dc30b72" containerID="bd2ba327ec5a6c52b21be001e26154a58c5fc191658a3eadcaafcc04b9d9a4aa" exitCode=0 Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.637555 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" event={"ID":"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72","Type":"ContainerDied","Data":"bd2ba327ec5a6c52b21be001e26154a58c5fc191658a3eadcaafcc04b9d9a4aa"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.637601 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.643418 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mnl8k"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.643771 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-7lswj" podStartSLOduration=118.643754798 podStartE2EDuration="1m58.643754798s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:41.639820778 +0000 UTC m=+142.582113808" watchObservedRunningTime="2025-12-15 12:19:41.643754798 +0000 UTC m=+142.586047828" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.686682 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:41 crc kubenswrapper[4719]: E1215 12:19:41.687949 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:42.187935496 +0000 UTC m=+143.130228526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.691743 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" event={"ID":"556c5cc5-d57e-413d-aa79-421d196cb799","Type":"ContainerStarted","Data":"2082cbdee92bbf92f31b7999981f3ea9b4a566f877d23e4329895c981ce1d428"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.692561 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.698651 4719 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-t6qkh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.698701 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.707462 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" event={"ID":"8a457b5d-1084-4330-bcf5-bd5c5ef97c85","Type":"ContainerStarted","Data":"fc7cfb5997180ea782bc2a4e6ff1410534903a426718592923d537f836bbdda6"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.707503 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" event={"ID":"8a457b5d-1084-4330-bcf5-bd5c5ef97c85","Type":"ContainerStarted","Data":"0d7cfe81930a4bf9acac018b1eacfbcf6e1453cb248fe27648a18e0815a03ab8"} Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.748914 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" podStartSLOduration=118.748896261 podStartE2EDuration="1m58.748896261s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:41.746987567 +0000 UTC m=+142.689280587" watchObservedRunningTime="2025-12-15 12:19:41.748896261 +0000 UTC m=+142.691189291" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.778461 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787"] Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.801823 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.802312 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" podStartSLOduration=118.802297875 podStartE2EDuration="1m58.802297875s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:41.800166073 +0000 UTC m=+142.742459103" watchObservedRunningTime="2025-12-15 12:19:41.802297875 +0000 UTC m=+142.744590905" Dec 15 12:19:41 crc kubenswrapper[4719]: E1215 12:19:41.809115 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:42.30908921 +0000 UTC m=+143.251382340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.855004 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2l8f2" podStartSLOduration=118.854985695 podStartE2EDuration="1m58.854985695s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:41.851437936 +0000 UTC m=+142.793730966" watchObservedRunningTime="2025-12-15 12:19:41.854985695 +0000 UTC m=+142.797278715" Dec 15 12:19:41 crc kubenswrapper[4719]: I1215 12:19:41.904804 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:41 crc kubenswrapper[4719]: E1215 12:19:41.905161 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:42.405149951 +0000 UTC m=+143.347442981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.005743 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:42 crc kubenswrapper[4719]: E1215 12:19:42.006218 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:42.506199357 +0000 UTC m=+143.448492387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.107515 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:42 crc kubenswrapper[4719]: E1215 12:19:42.107946 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:42.607932156 +0000 UTC m=+143.550225186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.208649 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:42 crc kubenswrapper[4719]: E1215 12:19:42.209404 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:42.709386396 +0000 UTC m=+143.651679426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.236282 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.310784 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:42 crc kubenswrapper[4719]: E1215 12:19:42.312457 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:42.812443819 +0000 UTC m=+143.754736859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.334393 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.414683 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:42 crc kubenswrapper[4719]: E1215 12:19:42.415056 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:42.915041588 +0000 UTC m=+143.857334618 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.516730 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:42 crc kubenswrapper[4719]: E1215 12:19:42.517229 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.017207181 +0000 UTC m=+143.959500221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.619352 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:42 crc kubenswrapper[4719]: E1215 12:19:42.619720 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.119705816 +0000 UTC m=+144.061998846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.637469 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:42 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:42 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:42 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.637522 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.730539 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:42 crc kubenswrapper[4719]: E1215 12:19:42.731125 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.231112677 +0000 UTC m=+144.173405707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.815764 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" event={"ID":"f2558ee2-4e5f-419a-94ca-5b37d181f83c","Type":"ContainerStarted","Data":"d8081f3e7705ecc5fed796695bba1817a3c045d4e06ae89aaaba00d4a6727d0b"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.834660 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:42 crc kubenswrapper[4719]: E1215 12:19:42.835605 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.335589667 +0000 UTC m=+144.277882697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.850378 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" event={"ID":"b03b5a84-a26a-4564-9081-3cfa516c02b3","Type":"ContainerStarted","Data":"de1773e5f9e9bccb08a26f3d1c46b86328ed3307a03c89a025e3f3c84eac501f"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.852323 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-vchb6" event={"ID":"fc14ce9a-a4fe-4be7-9fdd-e6e75e415829","Type":"ContainerStarted","Data":"e9ede444dcccfef44497bb41bab8049e40fc6da346bc5739a292b9d7e56e0365"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.852360 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-vchb6" event={"ID":"fc14ce9a-a4fe-4be7-9fdd-e6e75e415829","Type":"ContainerStarted","Data":"e6b361fd902b65941dd1bc0368036b3ebe71dae4714de5cb6c90e24ecd027600"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.875752 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zcgr5" podStartSLOduration=119.87573349 podStartE2EDuration="1m59.87573349s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:42.874363885 +0000 UTC m=+143.816656925" watchObservedRunningTime="2025-12-15 12:19:42.87573349 +0000 UTC m=+143.818026520" Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.918527 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" event={"ID":"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e","Type":"ContainerStarted","Data":"2ce8c6869c498393b00426927703a524e9bb87e9be56adda2d9025c654cc509e"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.918572 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" event={"ID":"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e","Type":"ContainerStarted","Data":"995387ebeb1472ce3f4773639977a441355852ffc4f47b7ba24cbe8720de3a05"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.922182 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" event={"ID":"e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1","Type":"ContainerStarted","Data":"f9ab6e66e768650beebf437a94180d156e9c36009d42e7385e119465b9b0d04b"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.923176 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-w28lp" event={"ID":"b44673b6-1b68-46e8-9b5f-b14c3389a017","Type":"ContainerStarted","Data":"f0cfaf206cd8aed376f3102c0bf3555a2e82d7d454cda48553e39353c007f930"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.933237 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fr486" podStartSLOduration=119.93322078 podStartE2EDuration="1m59.93322078s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:42.933139337 +0000 UTC m=+143.875432367" watchObservedRunningTime="2025-12-15 12:19:42.93322078 +0000 UTC m=+143.875513810" Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.937223 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" event={"ID":"d620f9fd-1374-4f93-972d-93f2b9173471","Type":"ContainerStarted","Data":"f3db6afbbcd0d21c1e6649a41ec1e66fada7a3ecc866436ac88b6892ac9721f9"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.939479 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:42 crc kubenswrapper[4719]: E1215 12:19:42.939767 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.439754796 +0000 UTC m=+144.382047826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.948759 4719 generic.go:334] "Generic (PLEG): container finished" podID="8f2d8d73-a1cc-4a67-9006-205567f3b907" containerID="1d36915ad483ccafc1cec308c3c7ce135d58029cd5131d502db8d492e6baf563" exitCode=0 Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.949056 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" event={"ID":"8f2d8d73-a1cc-4a67-9006-205567f3b907","Type":"ContainerStarted","Data":"c534049ac6630abce9c3692db46104ab1f1eda8d12981d0c3ecd6f1f5524468a"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.949081 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" event={"ID":"8f2d8d73-a1cc-4a67-9006-205567f3b907","Type":"ContainerDied","Data":"1d36915ad483ccafc1cec308c3c7ce135d58029cd5131d502db8d492e6baf563"} Dec 15 12:19:42 crc kubenswrapper[4719]: I1215 12:19:42.949599 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.011579 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-7lswj" event={"ID":"3a91fe95-e935-4e7c-bee1-d3d291206f67","Type":"ContainerStarted","Data":"a101c59dd6d7bf2631bbcc9e7fdc0d74b3bca151780bd2017d0e8e8d5563f1ff"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.025419 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-vchb6" podStartSLOduration=8.025403412 podStartE2EDuration="8.025403412s" podCreationTimestamp="2025-12-15 12:19:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.023988334 +0000 UTC m=+143.966281364" watchObservedRunningTime="2025-12-15 12:19:43.025403412 +0000 UTC m=+143.967696442" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.040162 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:43 crc kubenswrapper[4719]: E1215 12:19:43.041550 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.541527847 +0000 UTC m=+144.483820917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.049673 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ls926" event={"ID":"1b3e2612-1a37-4784-940f-44cb98fa747a","Type":"ContainerStarted","Data":"3530ca3e79d268540d59a2bee44f17a765ff35592e6bba4856bc89b7bd8de3da"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.049718 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ls926" event={"ID":"1b3e2612-1a37-4784-940f-44cb98fa747a","Type":"ContainerStarted","Data":"b553ead4d8cddb9f226f583f8c641c76211443fbddd1409c78602d528cde2114"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.082706 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cdb5p" event={"ID":"129630ef-67df-4003-aead-ea250b93c8b5","Type":"ContainerStarted","Data":"dad8e76b5f46539bd4d3f0a658aa2e85d987993eb3aab0c0d5ea01a73cbe266f"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.101650 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" event={"ID":"556c5cc5-d57e-413d-aa79-421d196cb799","Type":"ContainerStarted","Data":"cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.102364 4719 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-t6qkh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.102410 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.113873 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" event={"ID":"bda713e9-91f5-40bc-806e-205f4a5dae25","Type":"ContainerStarted","Data":"aee7813018d3c21b2b55fae3d2b58f588b1980de4f84b27d6c31daf8749c08ff"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.124969 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2xxrl" event={"ID":"c8458395-1102-446b-8b75-752820a9cfdb","Type":"ContainerStarted","Data":"01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.125009 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2xxrl" event={"ID":"c8458395-1102-446b-8b75-752820a9cfdb","Type":"ContainerStarted","Data":"3d9bbd2d1af5dac6cdaf45afb1ed3c8f062c0588580ca88a3f1283273ec433b2"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.137931 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" event={"ID":"0dd6e306-1355-43dd-895f-598d3d873a42","Type":"ContainerStarted","Data":"cd1189baca23df383552d66223a9c6557f47c68e79159d45c2b059e2694c6ba0"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.184342 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:43 crc kubenswrapper[4719]: E1215 12:19:43.187321 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.68730338 +0000 UTC m=+144.629596410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.218556 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" event={"ID":"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51","Type":"ContainerStarted","Data":"43d07ac5794200f80d34778ca3d8a1a55200521f904a19c956caa9a3d7542aa4"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.219667 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.223311 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" podStartSLOduration=120.223277544 podStartE2EDuration="2m0.223277544s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.150611841 +0000 UTC m=+144.092904871" watchObservedRunningTime="2025-12-15 12:19:43.223277544 +0000 UTC m=+144.165570594" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.223500 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ngv2q" podStartSLOduration=120.223492502 podStartE2EDuration="2m0.223492502s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.077994029 +0000 UTC m=+144.020287059" watchObservedRunningTime="2025-12-15 12:19:43.223492502 +0000 UTC m=+144.165785542" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.234995 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-2xxrl" podStartSLOduration=120.234976543 podStartE2EDuration="2m0.234976543s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.204505171 +0000 UTC m=+144.146798201" watchObservedRunningTime="2025-12-15 12:19:43.234976543 +0000 UTC m=+144.177269573" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.250635 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-ls926" podStartSLOduration=8.250614792 podStartE2EDuration="8.250614792s" podCreationTimestamp="2025-12-15 12:19:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.244754637 +0000 UTC m=+144.187047667" watchObservedRunningTime="2025-12-15 12:19:43.250614792 +0000 UTC m=+144.192907822" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.269753 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" event={"ID":"32919d63-a33b-44ac-af4e-aeed2e97a3ac","Type":"ContainerStarted","Data":"9679412da3ede232ddf610a1eb956fd864fa1342ccb9cbf95059472dcbbbeb97"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.270777 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.285645 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:43 crc kubenswrapper[4719]: E1215 12:19:43.286104 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.786080191 +0000 UTC m=+144.728373221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.286620 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:43 crc kubenswrapper[4719]: E1215 12:19:43.292280 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.792260645 +0000 UTC m=+144.734553675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.313841 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-c5df6" podStartSLOduration=120.313797651 podStartE2EDuration="2m0.313797651s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.2836567 +0000 UTC m=+144.225949730" watchObservedRunningTime="2025-12-15 12:19:43.313797651 +0000 UTC m=+144.256090671" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.317369 4719 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-5gcx2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.317427 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" podUID="32919d63-a33b-44ac-af4e-aeed2e97a3ac" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.359036 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" event={"ID":"4fd3ac7a-862d-45ff-a931-5962b105ac6e","Type":"ContainerStarted","Data":"84fbbf82b0af6fcb83dca71d625ed6c22040f333c461bcca64e72cc69df6c3b5"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.359085 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" event={"ID":"4fd3ac7a-862d-45ff-a931-5962b105ac6e","Type":"ContainerStarted","Data":"c5cd1c68095b528067692fee7f2f0d40e1b586a1d4e3f149db104398a36a7c0b"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.389879 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:43 crc kubenswrapper[4719]: E1215 12:19:43.392164 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.892145854 +0000 UTC m=+144.834438884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.392735 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" event={"ID":"4d7d0aab-ec25-437b-99e5-4a39a14bac0c","Type":"ContainerStarted","Data":"77c7ac32f068d516f680b95ee478c62edbd36951949a91b5f8b31e329cb34662"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.392762 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" podStartSLOduration=120.392752604 podStartE2EDuration="2m0.392752604s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.391328556 +0000 UTC m=+144.333621586" watchObservedRunningTime="2025-12-15 12:19:43.392752604 +0000 UTC m=+144.335045634" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.410152 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-glf8c" event={"ID":"daadac3d-d689-4424-a6ff-1a6da646d03f","Type":"ContainerStarted","Data":"333718a4e7d746aa43f203bf274cb3b602c410e51006f077a6de3740049aa480"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.411502 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-glf8c" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.413427 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" event={"ID":"273927d9-dcea-4e2c-8c0c-c80cb6902cda","Type":"ContainerStarted","Data":"f0e391be48964945c561372c0bf554effa9ff02a0ad4d8b5dc79efd73b9fdc63"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.415518 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" event={"ID":"4cec2f60-1603-4a04-baab-3d9c070257da","Type":"ContainerStarted","Data":"c50f57b47a870b172a15ce699c99887afc1e464de0ed360e4ce754be45841d53"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.415543 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" event={"ID":"4cec2f60-1603-4a04-baab-3d9c070257da","Type":"ContainerStarted","Data":"3ffc0d2b5a25ab68bd27392df2dbe1662db5b1d9facb863825dca24a07ba1a31"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.432701 4719 patch_prober.go:28] interesting pod/downloads-7954f5f757-glf8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.432757 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-glf8c" podUID="daadac3d-d689-4424-a6ff-1a6da646d03f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.477490 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" podStartSLOduration=120.477469318 podStartE2EDuration="2m0.477469318s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.477041353 +0000 UTC m=+144.419334393" watchObservedRunningTime="2025-12-15 12:19:43.477469318 +0000 UTC m=+144.419762338" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.489150 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" event={"ID":"790b2072-3dd0-4e6c-896c-724a3cd2bc5d","Type":"ContainerStarted","Data":"b5faf357ba7cfe73fec1459f4410786281431fe0ecdb13f62fde96d9481c7c10"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.489196 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.492219 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:43 crc kubenswrapper[4719]: E1215 12:19:43.493302 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:43.993291034 +0000 UTC m=+144.935584064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.511314 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-827sb" event={"ID":"392f2574-b01c-41bf-b0df-7774696560ee","Type":"ContainerStarted","Data":"295f4014829784fce64a92626a851b9bf021aae7be09e27328a650f05d80b1c6"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.511355 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-827sb" event={"ID":"392f2574-b01c-41bf-b0df-7774696560ee","Type":"ContainerStarted","Data":"bbe11031258fbfa3199b2711d6980b4d9af033812cb59ab31bc339359cbdc9b0"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.513644 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-glf8c" podStartSLOduration=120.513625918 podStartE2EDuration="2m0.513625918s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.513078931 +0000 UTC m=+144.455371951" watchObservedRunningTime="2025-12-15 12:19:43.513625918 +0000 UTC m=+144.455918948" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.522069 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" event={"ID":"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4","Type":"ContainerStarted","Data":"458b490a83258990066412fcd4d7df3d388b5d4083fc4cd8ff350e1431a92bd4"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.522969 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.533761 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp" event={"ID":"94ce3699-c8e3-42df-bafe-8cdd7572c406","Type":"ContainerStarted","Data":"3aca46c3a4125e456106da42fadb1a23b79f456d17b40aba1c36669ee0547ef4"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.542046 4719 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-txqnw container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.542155 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" podUID="674241e5-eef3-4cf8-b4ff-c14a7a20d1b4" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.543929 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" event={"ID":"a2ade101-54a5-4915-9b0c-bc668fe625ff","Type":"ContainerStarted","Data":"c9f0bb9f5910498961851c2eab5c8f71e85a96f7103bf3e88c880f363aca22e3"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.543975 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" event={"ID":"a2ade101-54a5-4915-9b0c-bc668fe625ff","Type":"ContainerStarted","Data":"2e2d34fed5818d5ff754ed58a798b02faf79913c6b4a54c3e1e4470cdc74751d"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.545062 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.589696 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" event={"ID":"994235ed-c7e5-44d3-bef7-c8df6b3224a3","Type":"ContainerStarted","Data":"255630a252de9ac3dc697cc533b693a93f65e5aab113c2341b9f7afd2b069d4a"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.596400 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:43 crc kubenswrapper[4719]: E1215 12:19:43.597051 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:44.09703562 +0000 UTC m=+145.039328650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.639449 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-p5n7r" podStartSLOduration=120.639430617 podStartE2EDuration="2m0.639430617s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.638405204 +0000 UTC m=+144.580698224" watchObservedRunningTime="2025-12-15 12:19:43.639430617 +0000 UTC m=+144.581723647" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.639960 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" event={"ID":"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0","Type":"ContainerStarted","Data":"c9899fb9ef1eee9a7d274b4b0a0b19efad1b52f69953ba194b9d1d48a1f82eea"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.639998 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" event={"ID":"d34646d2-f56c-4b6a-bbaa-c9f3485b61e0","Type":"ContainerStarted","Data":"20f522fc1b49cff79d05c22360bf60745a52d61d4b888d079325783a8402b1a8"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.642417 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:43 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:43 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:43 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.642460 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.660281 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.673386 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" event={"ID":"3a4fc82e-33f4-4b77-ad47-54554d0c08b4","Type":"ContainerStarted","Data":"06dff12728b1772a93cc24715ebebd790d4d945e61303f06f67a6b530eb848a4"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.698508 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:43 crc kubenswrapper[4719]: E1215 12:19:43.700232 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:44.200218047 +0000 UTC m=+145.142511077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.718571 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" event={"ID":"d561cf62-c184-4256-bec7-af48aa15afcb","Type":"ContainerStarted","Data":"b2f9a9041fa37e8f0c2a7718132aa06a00df6b6776a143122c09bac44e396d36"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.742214 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" event={"ID":"1699adb2-2c5e-4345-92bc-184205052fc8","Type":"ContainerStarted","Data":"fa2bfb43f72dc22e22b7c635f8a3833dbbe07f8922a324f24be92b468b146df2"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.749570 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7g9lh" podStartSLOduration=120.749553415 podStartE2EDuration="2m0.749553415s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.730551064 +0000 UTC m=+144.672844094" watchObservedRunningTime="2025-12-15 12:19:43.749553415 +0000 UTC m=+144.691846445" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.772941 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" event={"ID":"6b7161e6-9af7-4443-9af7-2d11e0f36d68","Type":"ContainerStarted","Data":"314c874c4e3152ac00de103537680b9ec1178b4aeed00021c61d0848f3c3ce81"} Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.804419 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.804804 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:43 crc kubenswrapper[4719]: E1215 12:19:43.805728 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:44.305713371 +0000 UTC m=+145.248006401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.830613 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lsvbn" podStartSLOduration=120.830600947 podStartE2EDuration="2m0.830600947s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:43.829948856 +0000 UTC m=+144.772241896" watchObservedRunningTime="2025-12-15 12:19:43.830600947 +0000 UTC m=+144.772893977" Dec 15 12:19:43 crc kubenswrapper[4719]: I1215 12:19:43.907651 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:43 crc kubenswrapper[4719]: E1215 12:19:43.910351 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:44.410337607 +0000 UTC m=+145.352630637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.015118 4719 patch_prober.go:28] interesting pod/console-operator-58897d9998-7lswj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.015420 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-7lswj" podUID="3a91fe95-e935-4e7c-bee1-d3d291206f67" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.015814 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.016114 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:44.516101249 +0000 UTC m=+145.458394279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.106088 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" podStartSLOduration=121.106068648 podStartE2EDuration="2m1.106068648s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:44.004105411 +0000 UTC m=+144.946398441" watchObservedRunningTime="2025-12-15 12:19:44.106068648 +0000 UTC m=+145.048361678" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.116974 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.117307 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:44.617294951 +0000 UTC m=+145.559587981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.218638 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.218778 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:44.718760211 +0000 UTC m=+145.661053241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.218834 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.219097 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:44.719090172 +0000 UTC m=+145.661383202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.220316 4719 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nkz2j container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.220354 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" podUID="4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.283747 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" podStartSLOduration=121.283727659 podStartE2EDuration="2m1.283727659s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:44.283199581 +0000 UTC m=+145.225492611" watchObservedRunningTime="2025-12-15 12:19:44.283727659 +0000 UTC m=+145.226020689" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.285351 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-94r2g" podStartSLOduration=121.285344753 podStartE2EDuration="2m1.285344753s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:44.117588871 +0000 UTC m=+145.059881901" watchObservedRunningTime="2025-12-15 12:19:44.285344753 +0000 UTC m=+145.227637783" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.319484 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.319892 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:44.819867849 +0000 UTC m=+145.762160879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.421284 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.421648 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:44.92163749 +0000 UTC m=+145.863930520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.522062 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.522455 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.022436758 +0000 UTC m=+145.964729788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.584247 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xqsdv" podStartSLOduration=121.5842165 podStartE2EDuration="2m1.5842165s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:44.442771371 +0000 UTC m=+145.385064401" watchObservedRunningTime="2025-12-15 12:19:44.5842165 +0000 UTC m=+145.526509530" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.623799 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.624099 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.124088265 +0000 UTC m=+146.066381295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.632463 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:44 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:44 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:44 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.632506 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.725479 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.725765 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.225750531 +0000 UTC m=+146.168043561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.799584 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" event={"ID":"273927d9-dcea-4e2c-8c0c-c80cb6902cda","Type":"ContainerStarted","Data":"10cef893e2b1ba628d4f5d74de0af52d49b158fb0fc8b663c32b7b3a1605d7ca"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.811321 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp" event={"ID":"94ce3699-c8e3-42df-bafe-8cdd7572c406","Type":"ContainerStarted","Data":"22206d6728a0629134f3c3c6e020da16dd4b0cf497c8339ca3109f9aa6cedec6"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.811360 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp" event={"ID":"94ce3699-c8e3-42df-bafe-8cdd7572c406","Type":"ContainerStarted","Data":"62d07f5aa27464b20175fac8406587835eb2f271c5d4822269b2e2a5a322b399"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.814115 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" event={"ID":"aaa5a8e9-a708-4115-95d6-0e1a68ef2b6e","Type":"ContainerStarted","Data":"ca0568c9225dab8cd79285762e060dea92f14fbeb6c0b8fd90c2b2bb10d29cd5"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.829637 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.830048 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.330034585 +0000 UTC m=+146.272327615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.832809 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" event={"ID":"32919d63-a33b-44ac-af4e-aeed2e97a3ac","Type":"ContainerStarted","Data":"5c66b2159a4198db77252936f80379cd31e070d3fb2c021296a699678aa72a00"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.835928 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-827sb" event={"ID":"392f2574-b01c-41bf-b0df-7774696560ee","Type":"ContainerStarted","Data":"ab76f5e9182d9b9a429b4925608a6891532c0555b115086dd1ca0a2d56172fca"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.836298 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-827sb" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.837315 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" event={"ID":"d561cf62-c184-4256-bec7-af48aa15afcb","Type":"ContainerStarted","Data":"884550579c4dcb86fd4f0f0fcfb655c05594c73113f104382b9d98efec8ba007"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.837337 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" event={"ID":"d561cf62-c184-4256-bec7-af48aa15afcb","Type":"ContainerStarted","Data":"c5a0ad2a55a041c3e6ff8070fabf60b657b80b0fed7ac8d58968cf56a7fe1cb2"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.838681 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" podStartSLOduration=121.838671362 podStartE2EDuration="2m1.838671362s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:44.838291909 +0000 UTC m=+145.780584969" watchObservedRunningTime="2025-12-15 12:19:44.838671362 +0000 UTC m=+145.780964392" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.850327 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" event={"ID":"30f62441-ac5d-48b7-b77c-8b6ff58546d1","Type":"ContainerStarted","Data":"eb38c907338459bcb2f5421f4e7f966c897e86c316a0d0c6de75d741fcbeb316"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.850387 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" event={"ID":"30f62441-ac5d-48b7-b77c-8b6ff58546d1","Type":"ContainerStarted","Data":"bb60ef139ecbcf7fe4e72d3faae433049cc2e7d524211d1cfb34eb153a8c0925"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.851980 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" event={"ID":"bda713e9-91f5-40bc-806e-205f4a5dae25","Type":"ContainerStarted","Data":"3c667eaf6f0c5969a60eb98396fe1fa31800ae513e024823226fa09a49501910"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.852004 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" event={"ID":"bda713e9-91f5-40bc-806e-205f4a5dae25","Type":"ContainerStarted","Data":"9843dbc8ad3ae27d6d4f4d75c050c1b7c1c1ae7dbcb3219a53ec973dc64c608b"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.870138 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" event={"ID":"3a4fc82e-33f4-4b77-ad47-54554d0c08b4","Type":"ContainerStarted","Data":"a93a67b631d815390a4a1a9e42bb6a0da0e48156c35c207f11fd70d342be562e"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.871675 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" event={"ID":"674241e5-eef3-4cf8-b4ff-c14a7a20d1b4","Type":"ContainerStarted","Data":"4ed60986a7ea6c617aa811da8b5050457133724a52cc01c9d54733a1e50679af"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.876205 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ts787" event={"ID":"1699adb2-2c5e-4345-92bc-184205052fc8","Type":"ContainerStarted","Data":"7b131e6e21a7117e90c471824cda029e750542d2de3a558c705a0482404e0756"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.890218 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-txqnw" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.890258 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" event={"ID":"e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1","Type":"ContainerStarted","Data":"b7ef05e5807ffc6f8f982c7bc46bc703728671e11c11935036d549f51cf7d296"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.890276 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" event={"ID":"e2afb3da-5a0b-48a1-8d8d-3e3aa3ecf6a1","Type":"ContainerStarted","Data":"913ca9ce50fd2f8d5630ab12ef7c2f1748f87a351bac24c52509660d9694dd89"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.902645 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" event={"ID":"790b2072-3dd0-4e6c-896c-724a3cd2bc5d","Type":"ContainerStarted","Data":"db4250e55875a62acbeee4bf1f5cc4c7c48a205cb287fd6dac360f31cc9f3b26"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.914368 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" event={"ID":"8b58d079-d1d0-4923-8ba3-bfeb1dc30b72","Type":"ContainerStarted","Data":"97aaaa6af68700f04a6e1f4762f7971d9e717f7e9a5fa8e18fb293f6d042fe57"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.921793 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" event={"ID":"6b7161e6-9af7-4443-9af7-2d11e0f36d68","Type":"ContainerStarted","Data":"8f35255c44e421ff3bc388351b808061dae661bf0528ca8ef18a6572409ba608"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.923701 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-z56dp" event={"ID":"994235ed-c7e5-44d3-bef7-c8df6b3224a3","Type":"ContainerStarted","Data":"a730cc924c067386ff10adebbe1fbc0996464eb964b4711bfbe43ba2d7780827"} Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.927961 4719 patch_prober.go:28] interesting pod/downloads-7954f5f757-glf8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.928002 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-glf8c" podUID="daadac3d-d689-4424-a6ff-1a6da646d03f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.928039 4719 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-t6qkh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.928090 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Dec 15 12:19:44 crc kubenswrapper[4719]: I1215 12:19:44.934995 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:44 crc kubenswrapper[4719]: E1215 12:19:44.943749 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.443728212 +0000 UTC m=+146.386021242 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.003675 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" podStartSLOduration=122.003652452 podStartE2EDuration="2m2.003652452s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:44.934479315 +0000 UTC m=+145.876772335" watchObservedRunningTime="2025-12-15 12:19:45.003652452 +0000 UTC m=+145.945945492" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.004035 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hxfsm" podStartSLOduration=122.004031225 podStartE2EDuration="2m2.004031225s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:45.002317288 +0000 UTC m=+145.944610328" watchObservedRunningTime="2025-12-15 12:19:45.004031225 +0000 UTC m=+145.946324255" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.028490 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-7lswj" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.037398 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.038411 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.538387676 +0000 UTC m=+146.480680706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.138510 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.138729 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.638688438 +0000 UTC m=+146.580981468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.138935 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.139230 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.639216436 +0000 UTC m=+146.581509466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.188826 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" podStartSLOduration=122.188808683 podStartE2EDuration="2m2.188808683s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:45.035081916 +0000 UTC m=+145.977374946" watchObservedRunningTime="2025-12-15 12:19:45.188808683 +0000 UTC m=+146.131101713" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.239927 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.240052 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.740035804 +0000 UTC m=+146.682328834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.240206 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.240516 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.74050874 +0000 UTC m=+146.682801770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.331569 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" podStartSLOduration=122.331550664 podStartE2EDuration="2m2.331550664s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:45.201753022 +0000 UTC m=+146.144046052" watchObservedRunningTime="2025-12-15 12:19:45.331550664 +0000 UTC m=+146.273843694" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.340719 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.340957 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.840928926 +0000 UTC m=+146.783221966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.341275 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.341541 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.841533405 +0000 UTC m=+146.783826435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.442304 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.442507 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.942490579 +0000 UTC m=+146.884783599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.442611 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.442901 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:45.942893982 +0000 UTC m=+146.885187012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.467803 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-827sb" podStartSLOduration=10.467788059 podStartE2EDuration="10.467788059s" podCreationTimestamp="2025-12-15 12:19:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:45.467706066 +0000 UTC m=+146.409999096" watchObservedRunningTime="2025-12-15 12:19:45.467788059 +0000 UTC m=+146.410081089" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.543244 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.543435 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.04338893 +0000 UTC m=+146.985681960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.543469 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.543827 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.043819095 +0000 UTC m=+146.986112125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.635339 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:45 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:45 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:45 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.635397 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.644468 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.644743 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.144728007 +0000 UTC m=+147.087021037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.748002 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zhwtq" podStartSLOduration=122.747987686 podStartE2EDuration="2m2.747987686s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:45.657647296 +0000 UTC m=+146.599940326" watchObservedRunningTime="2025-12-15 12:19:45.747987686 +0000 UTC m=+146.690280716" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.748524 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.748602 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4cwjp" podStartSLOduration=122.748597167 podStartE2EDuration="2m2.748597167s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:45.747314635 +0000 UTC m=+146.689607665" watchObservedRunningTime="2025-12-15 12:19:45.748597167 +0000 UTC m=+146.690890197" Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.748814 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.248803094 +0000 UTC m=+147.191096114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.765371 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.834606 4719 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-5gcx2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.834685 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" podUID="32919d63-a33b-44ac-af4e-aeed2e97a3ac" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.849517 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.849917 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.34983707 +0000 UTC m=+147.292130110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.920745 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-9twm9" podStartSLOduration=122.920726215 podStartE2EDuration="2m2.920726215s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:45.866285436 +0000 UTC m=+146.808578496" watchObservedRunningTime="2025-12-15 12:19:45.920726215 +0000 UTC m=+146.863019245" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.929825 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" event={"ID":"273927d9-dcea-4e2c-8c0c-c80cb6902cda","Type":"ContainerStarted","Data":"30a037802da428a2af223b838bcea771d595b9b81c70af3c4d17d4f01f16a4e1"} Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.929961 4719 patch_prober.go:28] interesting pod/downloads-7954f5f757-glf8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.929997 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-glf8c" podUID="daadac3d-d689-4424-a6ff-1a6da646d03f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.950706 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:45 crc kubenswrapper[4719]: E1215 12:19:45.951101 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.451084313 +0000 UTC m=+147.393377343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.994389 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-mnl8k" podStartSLOduration=122.99437403 podStartE2EDuration="2m2.99437403s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:45.923466906 +0000 UTC m=+146.865759936" watchObservedRunningTime="2025-12-15 12:19:45.99437403 +0000 UTC m=+146.936667060" Dec 15 12:19:45 crc kubenswrapper[4719]: I1215 12:19:45.994548 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7z22d" podStartSLOduration=122.994542846 podStartE2EDuration="2m2.994542846s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:45.991014158 +0000 UTC m=+146.933307188" watchObservedRunningTime="2025-12-15 12:19:45.994542846 +0000 UTC m=+146.936835866" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.051741 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.053688 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.55367031 +0000 UTC m=+147.495963340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.156073 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.156382 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.656366941 +0000 UTC m=+147.598659961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.175764 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" podStartSLOduration=123.175723454 podStartE2EDuration="2m3.175723454s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:46.08407229 +0000 UTC m=+147.026365320" watchObservedRunningTime="2025-12-15 12:19:46.175723454 +0000 UTC m=+147.118016504" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.260187 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.260588 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.760573002 +0000 UTC m=+147.702866032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.361997 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.362394 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.862381814 +0000 UTC m=+147.804674834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.463024 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.463236 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.963208243 +0000 UTC m=+147.905501273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.463563 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.463847 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:46.963840614 +0000 UTC m=+147.906133644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.564639 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.565188 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.065173759 +0000 UTC m=+148.007466779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.633371 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:46 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:46 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:46 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.633435 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.666263 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.666321 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.666345 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.666384 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.666410 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.666755 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.166735013 +0000 UTC m=+148.109028043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.667117 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.676708 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.688655 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.702636 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.706111 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.713746 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.722166 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.767409 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.767799 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.26778029 +0000 UTC m=+148.210073320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.880622 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.881120 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.381109625 +0000 UTC m=+148.323402655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.934946 4719 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-fbzqs container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.935007 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" podUID="8f2d8d73-a1cc-4a67-9006-205567f3b907" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.935457 4719 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-5gcx2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.935510 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" podUID="32919d63-a33b-44ac-af4e-aeed2e97a3ac" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.962615 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" event={"ID":"273927d9-dcea-4e2c-8c0c-c80cb6902cda","Type":"ContainerStarted","Data":"d77abe6a9411500a15956658a5838fdd036a3f2a3b707fe2bbdc83ca112b357e"} Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.982156 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.982371 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.482350377 +0000 UTC m=+148.424643407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:46 crc kubenswrapper[4719]: I1215 12:19:46.982492 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:46 crc kubenswrapper[4719]: E1215 12:19:46.982794 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.482781881 +0000 UTC m=+148.425074911 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.083573 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.085141 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.585119621 +0000 UTC m=+148.527412651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.094765 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vqndj"] Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.095620 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:47 crc kubenswrapper[4719]: W1215 12:19:47.101264 4719 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.101311 4719 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.136376 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k2xhs"] Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.137398 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:47 crc kubenswrapper[4719]: W1215 12:19:47.158131 4719 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.158166 4719 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.185404 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.185454 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-utilities\") pod \"certified-operators-vqndj\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.185480 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzjhp\" (UniqueName: \"kubernetes.io/projected/722c74db-78bf-49c2-8ff7-03cba8ea3afc-kube-api-access-lzjhp\") pod \"certified-operators-vqndj\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.185534 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-catalog-content\") pod \"certified-operators-vqndj\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.185795 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.685781795 +0000 UTC m=+148.628074825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.211402 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2xhs"] Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.287020 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.287190 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbjvv\" (UniqueName: \"kubernetes.io/projected/cdd254ce-5006-40e1-a7f8-0850d049242f-kube-api-access-xbjvv\") pod \"community-operators-k2xhs\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.287226 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-utilities\") pod \"community-operators-k2xhs\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.287256 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-utilities\") pod \"certified-operators-vqndj\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.287288 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzjhp\" (UniqueName: \"kubernetes.io/projected/722c74db-78bf-49c2-8ff7-03cba8ea3afc-kube-api-access-lzjhp\") pod \"certified-operators-vqndj\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.287340 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-catalog-content\") pod \"community-operators-k2xhs\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.287358 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-catalog-content\") pod \"certified-operators-vqndj\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.287757 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-catalog-content\") pod \"certified-operators-vqndj\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.287848 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.787833704 +0000 UTC m=+148.730126734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.288145 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-utilities\") pod \"certified-operators-vqndj\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.309930 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vqndj"] Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.323885 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5r6hk"] Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.324761 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.326078 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzjhp\" (UniqueName: \"kubernetes.io/projected/722c74db-78bf-49c2-8ff7-03cba8ea3afc-kube-api-access-lzjhp\") pod \"certified-operators-vqndj\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.364343 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5r6hk"] Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.388237 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-catalog-content\") pod \"community-operators-k2xhs\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.388596 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbjvv\" (UniqueName: \"kubernetes.io/projected/cdd254ce-5006-40e1-a7f8-0850d049242f-kube-api-access-xbjvv\") pod \"community-operators-k2xhs\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.388620 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.388639 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-utilities\") pod \"community-operators-k2xhs\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.388670 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-utilities\") pod \"certified-operators-5r6hk\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.388692 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25lx5\" (UniqueName: \"kubernetes.io/projected/f162837e-a473-4ee1-ac47-3caca3b12bbc-kube-api-access-25lx5\") pod \"certified-operators-5r6hk\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.388729 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-catalog-content\") pod \"certified-operators-5r6hk\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.389184 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-catalog-content\") pod \"community-operators-k2xhs\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.389894 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.889883674 +0000 UTC m=+148.832176694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.390243 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-utilities\") pod \"community-operators-k2xhs\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.436362 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z2htt"] Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.437259 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.467786 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbjvv\" (UniqueName: \"kubernetes.io/projected/cdd254ce-5006-40e1-a7f8-0850d049242f-kube-api-access-xbjvv\") pod \"community-operators-k2xhs\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.494096 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.494260 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25lx5\" (UniqueName: \"kubernetes.io/projected/f162837e-a473-4ee1-ac47-3caca3b12bbc-kube-api-access-25lx5\") pod \"certified-operators-5r6hk\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.494296 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-catalog-content\") pod \"community-operators-z2htt\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.494570 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.99454551 +0000 UTC m=+148.936838560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.494642 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-utilities\") pod \"community-operators-z2htt\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.494730 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvnd5\" (UniqueName: \"kubernetes.io/projected/b166625d-5aa1-46c1-9ec7-54aec2503fb7-kube-api-access-qvnd5\") pod \"community-operators-z2htt\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.494773 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-catalog-content\") pod \"certified-operators-5r6hk\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.494916 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.494965 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-utilities\") pod \"certified-operators-5r6hk\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.495349 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-catalog-content\") pod \"certified-operators-5r6hk\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.495566 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:47.995557984 +0000 UTC m=+148.937851014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.495773 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-utilities\") pod \"certified-operators-5r6hk\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.525489 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.525533 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.558992 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.559847 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.579416 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z2htt"] Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.596783 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.597049 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-utilities\") pod \"community-operators-z2htt\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.597108 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvnd5\" (UniqueName: \"kubernetes.io/projected/b166625d-5aa1-46c1-9ec7-54aec2503fb7-kube-api-access-qvnd5\") pod \"community-operators-z2htt\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.597249 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-catalog-content\") pod \"community-operators-z2htt\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.597619 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-catalog-content\") pod \"community-operators-z2htt\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.597682 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.097667556 +0000 UTC m=+149.039960576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.597885 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-utilities\") pod \"community-operators-z2htt\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.600056 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.603593 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25lx5\" (UniqueName: \"kubernetes.io/projected/f162837e-a473-4ee1-ac47-3caca3b12bbc-kube-api-access-25lx5\") pod \"certified-operators-5r6hk\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.645308 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvnd5\" (UniqueName: \"kubernetes.io/projected/b166625d-5aa1-46c1-9ec7-54aec2503fb7-kube-api-access-qvnd5\") pod \"community-operators-z2htt\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.652380 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:47 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:47 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:47 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.652432 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.698576 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.699852 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.199840669 +0000 UTC m=+149.142133699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.716181 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbzqs" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.799492 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.799645 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.299623153 +0000 UTC m=+149.241916183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.799958 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.800353 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.300341158 +0000 UTC m=+149.242634188 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.805317 4719 patch_prober.go:28] interesting pod/apiserver-76f77b778f-68lw2 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]log ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]etcd ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]poststarthook/generic-apiserver-start-informers ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]poststarthook/max-in-flight-filter ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 15 12:19:47 crc kubenswrapper[4719]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 15 12:19:47 crc kubenswrapper[4719]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 15 12:19:47 crc kubenswrapper[4719]: [+]poststarthook/project.openshift.io-projectcache ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]poststarthook/openshift.io-startinformers ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 15 12:19:47 crc kubenswrapper[4719]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 15 12:19:47 crc kubenswrapper[4719]: livez check failed Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.805363 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" podUID="30f62441-ac5d-48b7-b77c-8b6ff58546d1" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.900883 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.901063 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.401046093 +0000 UTC m=+149.343339123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.901101 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:47 crc kubenswrapper[4719]: E1215 12:19:47.901423 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.401414535 +0000 UTC m=+149.343707565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.968425 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" event={"ID":"273927d9-dcea-4e2c-8c0c-c80cb6902cda","Type":"ContainerStarted","Data":"0bb550fe7469d065dfee083e0f6141ecbcf9f6d191e354e63d10c123acf13d8e"} Dec 15 12:19:47 crc kubenswrapper[4719]: I1215 12:19:47.992230 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gp49f" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.003236 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:48 crc kubenswrapper[4719]: E1215 12:19:48.003587 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.503572249 +0000 UTC m=+149.445865279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.030011 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-4mbgz" podStartSLOduration=13.029993806 podStartE2EDuration="13.029993806s" podCreationTimestamp="2025-12-15 12:19:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:48.028329021 +0000 UTC m=+148.970622051" watchObservedRunningTime="2025-12-15 12:19:48.029993806 +0000 UTC m=+148.972286826" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.104552 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:48 crc kubenswrapper[4719]: E1215 12:19:48.105327 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.605313197 +0000 UTC m=+149.547606307 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.207004 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:48 crc kubenswrapper[4719]: E1215 12:19:48.207663 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.707647577 +0000 UTC m=+149.649940607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.309601 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:48 crc kubenswrapper[4719]: E1215 12:19:48.310742 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.810730241 +0000 UTC m=+149.753023271 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.351824 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.352208 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.360061 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.412267 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:48 crc kubenswrapper[4719]: E1215 12:19:48.412715 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:48.912698558 +0000 UTC m=+149.854991578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.424476 4719 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/certified-operators-vqndj" secret="" err="failed to sync secret cache: timed out waiting for the condition" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.424569 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.498371 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.503531 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.513667 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:48 crc kubenswrapper[4719]: E1215 12:19:48.514165 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.014145907 +0000 UTC m=+149.956438937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.614981 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:48 crc kubenswrapper[4719]: E1215 12:19:48.615366 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.115351699 +0000 UTC m=+150.057644729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.628534 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.631642 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:48 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:48 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:48 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.631726 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.717224 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:48 crc kubenswrapper[4719]: E1215 12:19:48.717563 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.217547804 +0000 UTC m=+150.159840834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.814124 4719 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.820429 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:48 crc kubenswrapper[4719]: E1215 12:19:48.820896 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.320881017 +0000 UTC m=+150.263174047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.921646 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:48 crc kubenswrapper[4719]: E1215 12:19:48.922197 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.42216044 +0000 UTC m=+150.364453470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.983346 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"51dc04e4d57c3a3524b860aa82f27d75e397ea1aaa3c0be560cfc601ee4021f4"} Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.983397 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cbe7c36b54be6e598998046e215fb61cb2a35bdec456b824eeee785cdf5bc64f"} Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.984158 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.992250 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"57fbad1093bd994e64796b304b8f97c7ba448cae54821000a5b80b241fd1cce0"} Dec 15 12:19:48 crc kubenswrapper[4719]: I1215 12:19:48.992316 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c0b16856aeb47b17d2725cd3544724109ce0107dfa7a078b8e88f203dde8f676"} Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.008271 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f49ab0aa4b39f5616e07bbb11d0e84361f0d91f708c6b98a88815ec4480ff1ab"} Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.008308 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9d9dc01c3179bfc03c5f10c3251c0ba69a31c5bf15eec618ff8711f72552c289"} Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.022411 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:49 crc kubenswrapper[4719]: E1215 12:19:49.022878 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.522840335 +0000 UTC m=+150.465133365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.059707 4719 patch_prober.go:28] interesting pod/downloads-7954f5f757-glf8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.059787 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-glf8c" podUID="daadac3d-d689-4424-a6ff-1a6da646d03f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.059980 4719 patch_prober.go:28] interesting pod/downloads-7954f5f757-glf8c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.060033 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-glf8c" podUID="daadac3d-d689-4424-a6ff-1a6da646d03f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.072957 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.124460 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:49 crc kubenswrapper[4719]: E1215 12:19:49.125496 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.625468564 +0000 UTC m=+150.567761594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.200565 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.217122 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.217204 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5gcx2" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.229633 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:49 crc kubenswrapper[4719]: E1215 12:19:49.231134 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.731118723 +0000 UTC m=+150.673411753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.244179 4719 patch_prober.go:28] interesting pod/console-f9d7485db-2xxrl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.244273 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2xxrl" podUID="c8458395-1102-446b-8b75-752820a9cfdb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.248586 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kh6zf"] Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.265922 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.282495 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.286249 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kh6zf"] Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.331690 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbdq8\" (UniqueName: \"kubernetes.io/projected/20547bc3-5b4d-40fd-898e-49b148705327-kube-api-access-cbdq8\") pod \"redhat-marketplace-kh6zf\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.331734 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-utilities\") pod \"redhat-marketplace-kh6zf\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.331754 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-catalog-content\") pod \"redhat-marketplace-kh6zf\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.331814 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:49 crc kubenswrapper[4719]: E1215 12:19:49.334142 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.834131665 +0000 UTC m=+150.776424695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.435314 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:49 crc kubenswrapper[4719]: E1215 12:19:49.435417 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.935400969 +0000 UTC m=+150.877693999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.435643 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:49 crc kubenswrapper[4719]: E1215 12:19:49.435918 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:49.935911626 +0000 UTC m=+150.878204656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.436114 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbdq8\" (UniqueName: \"kubernetes.io/projected/20547bc3-5b4d-40fd-898e-49b148705327-kube-api-access-cbdq8\") pod \"redhat-marketplace-kh6zf\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.436163 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-utilities\") pod \"redhat-marketplace-kh6zf\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.436186 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-catalog-content\") pod \"redhat-marketplace-kh6zf\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.436643 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-catalog-content\") pod \"redhat-marketplace-kh6zf\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.437201 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-utilities\") pod \"redhat-marketplace-kh6zf\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.501431 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k2xhs"] Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.518561 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbdq8\" (UniqueName: \"kubernetes.io/projected/20547bc3-5b4d-40fd-898e-49b148705327-kube-api-access-cbdq8\") pod \"redhat-marketplace-kh6zf\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.540406 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:49 crc kubenswrapper[4719]: E1215 12:19:49.540750 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:50.040734088 +0000 UTC m=+150.983027118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.609769 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.644370 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:49 crc kubenswrapper[4719]: E1215 12:19:49.645235 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-15 12:19:50.145222698 +0000 UTC m=+151.087515728 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gfmg4" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.651068 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:49 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:49 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:49 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.651117 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.674550 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ckh9g"] Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.675748 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.728249 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z2htt"] Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.743896 4719 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-15T12:19:48.81436357Z","Handler":null,"Name":""} Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.745517 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.745725 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n78w\" (UniqueName: \"kubernetes.io/projected/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-kube-api-access-5n78w\") pod \"redhat-marketplace-ckh9g\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.745781 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-utilities\") pod \"redhat-marketplace-ckh9g\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.745833 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-catalog-content\") pod \"redhat-marketplace-ckh9g\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:49 crc kubenswrapper[4719]: E1215 12:19:49.753056 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-15 12:19:50.25303096 +0000 UTC m=+151.195323990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.761448 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckh9g"] Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.795501 4719 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.795542 4719 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.819112 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vqndj"] Dec 15 12:19:49 crc kubenswrapper[4719]: W1215 12:19:49.838325 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod722c74db_78bf_49c2_8ff7_03cba8ea3afc.slice/crio-c1761e87867d14dea8fbbc05f4a38dd13d99cd5e6bcf4562061db1f861bc0ba5 WatchSource:0}: Error finding container c1761e87867d14dea8fbbc05f4a38dd13d99cd5e6bcf4562061db1f861bc0ba5: Status 404 returned error can't find the container with id c1761e87867d14dea8fbbc05f4a38dd13d99cd5e6bcf4562061db1f861bc0ba5 Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.864180 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-catalog-content\") pod \"redhat-marketplace-ckh9g\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.864277 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n78w\" (UniqueName: \"kubernetes.io/projected/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-kube-api-access-5n78w\") pod \"redhat-marketplace-ckh9g\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.864320 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.864373 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-utilities\") pod \"redhat-marketplace-ckh9g\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.864820 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-utilities\") pod \"redhat-marketplace-ckh9g\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.865064 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-catalog-content\") pod \"redhat-marketplace-ckh9g\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.889371 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.890126 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.899018 4719 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.899065 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.909970 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.910915 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.923059 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.946749 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n78w\" (UniqueName: \"kubernetes.io/projected/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-kube-api-access-5n78w\") pod \"redhat-marketplace-ckh9g\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.969062 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97dac188-c1d2-4256-8aed-d4519e718ecd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"97dac188-c1d2-4256-8aed-d4519e718ecd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 15 12:19:49 crc kubenswrapper[4719]: I1215 12:19:49.969152 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97dac188-c1d2-4256-8aed-d4519e718ecd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"97dac188-c1d2-4256-8aed-d4519e718ecd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.038222 4719 generic.go:334] "Generic (PLEG): container finished" podID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerID="56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84" exitCode=0 Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.038299 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2xhs" event={"ID":"cdd254ce-5006-40e1-a7f8-0850d049242f","Type":"ContainerDied","Data":"56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84"} Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.038326 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2xhs" event={"ID":"cdd254ce-5006-40e1-a7f8-0850d049242f","Type":"ContainerStarted","Data":"dd3f41b2b96471a1030d748e6f98950385bde3ecb089646e180092de0a0504e6"} Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.040147 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.050286 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.059346 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqndj" event={"ID":"722c74db-78bf-49c2-8ff7-03cba8ea3afc","Type":"ContainerStarted","Data":"c1761e87867d14dea8fbbc05f4a38dd13d99cd5e6bcf4562061db1f861bc0ba5"} Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.072552 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2htt" event={"ID":"b166625d-5aa1-46c1-9ec7-54aec2503fb7","Type":"ContainerStarted","Data":"df6eeffeb3558d26d92511ae98595ff563e4551c9540e4095e61fe1ce84ef017"} Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.079151 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gwjsp"] Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.080016 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97dac188-c1d2-4256-8aed-d4519e718ecd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"97dac188-c1d2-4256-8aed-d4519e718ecd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.080720 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97dac188-c1d2-4256-8aed-d4519e718ecd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"97dac188-c1d2-4256-8aed-d4519e718ecd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.081946 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97dac188-c1d2-4256-8aed-d4519e718ecd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"97dac188-c1d2-4256-8aed-d4519e718ecd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.094010 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.116207 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.127181 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gwjsp"] Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.173728 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97dac188-c1d2-4256-8aed-d4519e718ecd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"97dac188-c1d2-4256-8aed-d4519e718ecd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.182531 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-utilities\") pod \"redhat-operators-gwjsp\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.182587 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-catalog-content\") pod \"redhat-operators-gwjsp\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.182657 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvhtz\" (UniqueName: \"kubernetes.io/projected/638821b0-deb7-40ee-bad1-b552994b4598-kube-api-access-hvhtz\") pod \"redhat-operators-gwjsp\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.201394 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5r6hk"] Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.215414 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tmfbz"] Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.223171 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.235676 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tmfbz"] Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.279422 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gfmg4\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.288754 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hcz6\" (UniqueName: \"kubernetes.io/projected/6014e2e6-c315-4f83-9266-33239e0c43c4-kube-api-access-7hcz6\") pod \"redhat-operators-tmfbz\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.288805 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-utilities\") pod \"redhat-operators-gwjsp\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.288825 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-catalog-content\") pod \"redhat-operators-tmfbz\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.288866 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-catalog-content\") pod \"redhat-operators-gwjsp\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.288901 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvhtz\" (UniqueName: \"kubernetes.io/projected/638821b0-deb7-40ee-bad1-b552994b4598-kube-api-access-hvhtz\") pod \"redhat-operators-gwjsp\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.288919 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-utilities\") pod \"redhat-operators-tmfbz\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.289506 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-utilities\") pod \"redhat-operators-gwjsp\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.289730 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-catalog-content\") pod \"redhat-operators-gwjsp\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.299463 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.368772 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvhtz\" (UniqueName: \"kubernetes.io/projected/638821b0-deb7-40ee-bad1-b552994b4598-kube-api-access-hvhtz\") pod \"redhat-operators-gwjsp\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.393146 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.393782 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.394015 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-catalog-content\") pod \"redhat-operators-tmfbz\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.394065 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-utilities\") pod \"redhat-operators-tmfbz\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.394117 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hcz6\" (UniqueName: \"kubernetes.io/projected/6014e2e6-c315-4f83-9266-33239e0c43c4-kube-api-access-7hcz6\") pod \"redhat-operators-tmfbz\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.394806 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-catalog-content\") pod \"redhat-operators-tmfbz\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.395458 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-utilities\") pod \"redhat-operators-tmfbz\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.440550 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hcz6\" (UniqueName: \"kubernetes.io/projected/6014e2e6-c315-4f83-9266-33239e0c43c4-kube-api-access-7hcz6\") pod \"redhat-operators-tmfbz\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.456004 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.575445 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.581610 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.638741 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:50 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:50 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:50 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.638788 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.656780 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.657412 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.660048 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.660278 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.675975 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.707149 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d649d11f-b46b-4535-a07f-99089284ace8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d649d11f-b46b-4535-a07f-99089284ace8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.707299 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d649d11f-b46b-4535-a07f-99089284ace8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d649d11f-b46b-4535-a07f-99089284ace8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.749731 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kh6zf"] Dec 15 12:19:50 crc kubenswrapper[4719]: W1215 12:19:50.799210 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20547bc3_5b4d_40fd_898e_49b148705327.slice/crio-c75f6e32799777d9187f19d09dded38af1d7f485f274b54fcc58653deab58e00 WatchSource:0}: Error finding container c75f6e32799777d9187f19d09dded38af1d7f485f274b54fcc58653deab58e00: Status 404 returned error can't find the container with id c75f6e32799777d9187f19d09dded38af1d7f485f274b54fcc58653deab58e00 Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.808953 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d649d11f-b46b-4535-a07f-99089284ace8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d649d11f-b46b-4535-a07f-99089284ace8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.809043 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d649d11f-b46b-4535-a07f-99089284ace8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d649d11f-b46b-4535-a07f-99089284ace8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.809201 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d649d11f-b46b-4535-a07f-99089284ace8-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d649d11f-b46b-4535-a07f-99089284ace8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.845732 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d649d11f-b46b-4535-a07f-99089284ace8-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d649d11f-b46b-4535-a07f-99089284ace8\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 15 12:19:50 crc kubenswrapper[4719]: I1215 12:19:50.882057 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckh9g"] Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.027669 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tmfbz"] Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.123652 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.126081 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckh9g" event={"ID":"ea0b8649-9e2e-46d2-b9f0-5c51f155891e","Type":"ContainerStarted","Data":"d962fed45db1ba5c1d15f919342492d8fb6ba48f92b4b4bb7b5187a65260cbb6"} Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.126122 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckh9g" event={"ID":"ea0b8649-9e2e-46d2-b9f0-5c51f155891e","Type":"ContainerStarted","Data":"34b313a0f1617e695decdbcf5461e411a96ff3013285407be7cc6b8664e2a5c6"} Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.130925 4719 generic.go:334] "Generic (PLEG): container finished" podID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerID="9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c" exitCode=0 Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.131220 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqndj" event={"ID":"722c74db-78bf-49c2-8ff7-03cba8ea3afc","Type":"ContainerDied","Data":"9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c"} Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.133349 4719 generic.go:334] "Generic (PLEG): container finished" podID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerID="188aa5738cf207b7e96ceab78fbf7778d4a8d1d8b50d71248a6eb7b01a014d9e" exitCode=0 Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.133455 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r6hk" event={"ID":"f162837e-a473-4ee1-ac47-3caca3b12bbc","Type":"ContainerDied","Data":"188aa5738cf207b7e96ceab78fbf7778d4a8d1d8b50d71248a6eb7b01a014d9e"} Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.133494 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r6hk" event={"ID":"f162837e-a473-4ee1-ac47-3caca3b12bbc","Type":"ContainerStarted","Data":"88382695961da997124f15894e15ffcc84b891e37e8669db3e4e2f23c45366d9"} Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.138650 4719 generic.go:334] "Generic (PLEG): container finished" podID="6b7161e6-9af7-4443-9af7-2d11e0f36d68" containerID="8f35255c44e421ff3bc388351b808061dae661bf0528ca8ef18a6572409ba608" exitCode=0 Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.139908 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" event={"ID":"6b7161e6-9af7-4443-9af7-2d11e0f36d68","Type":"ContainerDied","Data":"8f35255c44e421ff3bc388351b808061dae661bf0528ca8ef18a6572409ba608"} Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.141642 4719 generic.go:334] "Generic (PLEG): container finished" podID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerID="2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19" exitCode=0 Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.141703 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2htt" event={"ID":"b166625d-5aa1-46c1-9ec7-54aec2503fb7","Type":"ContainerDied","Data":"2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19"} Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.146618 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmfbz" event={"ID":"6014e2e6-c315-4f83-9266-33239e0c43c4","Type":"ContainerStarted","Data":"20f7f70b96379ba9eac00aed1edfa82d1466e92f01448abfb3940efc75e5b844"} Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.151355 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kh6zf" event={"ID":"20547bc3-5b4d-40fd-898e-49b148705327","Type":"ContainerStarted","Data":"196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0"} Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.151423 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kh6zf" event={"ID":"20547bc3-5b4d-40fd-898e-49b148705327","Type":"ContainerStarted","Data":"c75f6e32799777d9187f19d09dded38af1d7f485f274b54fcc58653deab58e00"} Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.296733 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gfmg4"] Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.304220 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gwjsp"] Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.320039 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.356705 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.356763 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.483649 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.622003 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.634007 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:51 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:51 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:51 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:51 crc kubenswrapper[4719]: I1215 12:19:51.634091 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:51 crc kubenswrapper[4719]: W1215 12:19:51.641964 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd649d11f_b46b_4535_a07f_99089284ace8.slice/crio-bc9fcd8bc405785b48c6721de171f97850a230494c1b67601765f67e5056eece WatchSource:0}: Error finding container bc9fcd8bc405785b48c6721de171f97850a230494c1b67601765f67e5056eece: Status 404 returned error can't find the container with id bc9fcd8bc405785b48c6721de171f97850a230494c1b67601765f67e5056eece Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.160980 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"97dac188-c1d2-4256-8aed-d4519e718ecd","Type":"ContainerStarted","Data":"449e271bb75e932b6ce65fb72c24c9972a68d4c7b4300cf944345ed56c359173"} Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.161318 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"97dac188-c1d2-4256-8aed-d4519e718ecd","Type":"ContainerStarted","Data":"f49d9959bf5aa5687ed3a10e093c8ef84f0272b758a0108858415935a1204a07"} Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.165137 4719 generic.go:334] "Generic (PLEG): container finished" podID="20547bc3-5b4d-40fd-898e-49b148705327" containerID="196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0" exitCode=0 Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.165210 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kh6zf" event={"ID":"20547bc3-5b4d-40fd-898e-49b148705327","Type":"ContainerDied","Data":"196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0"} Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.188439 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.188419165 podStartE2EDuration="3.188419165s" podCreationTimestamp="2025-12-15 12:19:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:52.187471783 +0000 UTC m=+153.129764823" watchObservedRunningTime="2025-12-15 12:19:52.188419165 +0000 UTC m=+153.130712195" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.190141 4719 generic.go:334] "Generic (PLEG): container finished" podID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerID="d962fed45db1ba5c1d15f919342492d8fb6ba48f92b4b4bb7b5187a65260cbb6" exitCode=0 Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.190235 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckh9g" event={"ID":"ea0b8649-9e2e-46d2-b9f0-5c51f155891e","Type":"ContainerDied","Data":"d962fed45db1ba5c1d15f919342492d8fb6ba48f92b4b4bb7b5187a65260cbb6"} Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.191356 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d649d11f-b46b-4535-a07f-99089284ace8","Type":"ContainerStarted","Data":"bc9fcd8bc405785b48c6721de171f97850a230494c1b67601765f67e5056eece"} Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.204167 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" event={"ID":"27255302-b4b6-40dc-a5fd-f6e376b7e05d","Type":"ContainerStarted","Data":"c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124"} Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.204209 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" event={"ID":"27255302-b4b6-40dc-a5fd-f6e376b7e05d","Type":"ContainerStarted","Data":"42a20bd63897c11e3e862c545d5b7d458dd2701956a4a1c6be8ed4bddeb150d3"} Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.205010 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.229172 4719 generic.go:334] "Generic (PLEG): container finished" podID="638821b0-deb7-40ee-bad1-b552994b4598" containerID="c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a" exitCode=0 Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.229277 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gwjsp" event={"ID":"638821b0-deb7-40ee-bad1-b552994b4598","Type":"ContainerDied","Data":"c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a"} Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.229310 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gwjsp" event={"ID":"638821b0-deb7-40ee-bad1-b552994b4598","Type":"ContainerStarted","Data":"eeac450a182307fd5f5f3a9f17bece47aea0dd411f9bcd4c98cefe259ba80d2a"} Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.236053 4719 generic.go:334] "Generic (PLEG): container finished" podID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerID="d1f3956b0b9cb7e360724843365a9a58d4a7f80f700bed717185698b68a8fcc9" exitCode=0 Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.236833 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmfbz" event={"ID":"6014e2e6-c315-4f83-9266-33239e0c43c4","Type":"ContainerDied","Data":"d1f3956b0b9cb7e360724843365a9a58d4a7f80f700bed717185698b68a8fcc9"} Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.248082 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" podStartSLOduration=129.248063425 podStartE2EDuration="2m9.248063425s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:52.247329471 +0000 UTC m=+153.189622541" watchObservedRunningTime="2025-12-15 12:19:52.248063425 +0000 UTC m=+153.190356455" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.503303 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.507827 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-68lw2" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.635933 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:52 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:52 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:52 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.635980 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.675606 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.736146 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w88c6\" (UniqueName: \"kubernetes.io/projected/6b7161e6-9af7-4443-9af7-2d11e0f36d68-kube-api-access-w88c6\") pod \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.736544 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b7161e6-9af7-4443-9af7-2d11e0f36d68-secret-volume\") pod \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.736636 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b7161e6-9af7-4443-9af7-2d11e0f36d68-config-volume\") pod \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\" (UID: \"6b7161e6-9af7-4443-9af7-2d11e0f36d68\") " Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.737638 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7161e6-9af7-4443-9af7-2d11e0f36d68-config-volume" (OuterVolumeSpecName: "config-volume") pod "6b7161e6-9af7-4443-9af7-2d11e0f36d68" (UID: "6b7161e6-9af7-4443-9af7-2d11e0f36d68"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.761398 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b7161e6-9af7-4443-9af7-2d11e0f36d68-kube-api-access-w88c6" (OuterVolumeSpecName: "kube-api-access-w88c6") pod "6b7161e6-9af7-4443-9af7-2d11e0f36d68" (UID: "6b7161e6-9af7-4443-9af7-2d11e0f36d68"). InnerVolumeSpecName "kube-api-access-w88c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.761452 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b7161e6-9af7-4443-9af7-2d11e0f36d68-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6b7161e6-9af7-4443-9af7-2d11e0f36d68" (UID: "6b7161e6-9af7-4443-9af7-2d11e0f36d68"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.837819 4719 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b7161e6-9af7-4443-9af7-2d11e0f36d68-config-volume\") on node \"crc\" DevicePath \"\"" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.837878 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w88c6\" (UniqueName: \"kubernetes.io/projected/6b7161e6-9af7-4443-9af7-2d11e0f36d68-kube-api-access-w88c6\") on node \"crc\" DevicePath \"\"" Dec 15 12:19:52 crc kubenswrapper[4719]: I1215 12:19:52.837895 4719 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b7161e6-9af7-4443-9af7-2d11e0f36d68-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 15 12:19:53 crc kubenswrapper[4719]: I1215 12:19:53.248664 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" event={"ID":"6b7161e6-9af7-4443-9af7-2d11e0f36d68","Type":"ContainerDied","Data":"314c874c4e3152ac00de103537680b9ec1178b4aeed00021c61d0848f3c3ce81"} Dec 15 12:19:53 crc kubenswrapper[4719]: I1215 12:19:53.248700 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="314c874c4e3152ac00de103537680b9ec1178b4aeed00021c61d0848f3c3ce81" Dec 15 12:19:53 crc kubenswrapper[4719]: I1215 12:19:53.248643 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb" Dec 15 12:19:53 crc kubenswrapper[4719]: I1215 12:19:53.257694 4719 generic.go:334] "Generic (PLEG): container finished" podID="97dac188-c1d2-4256-8aed-d4519e718ecd" containerID="449e271bb75e932b6ce65fb72c24c9972a68d4c7b4300cf944345ed56c359173" exitCode=0 Dec 15 12:19:53 crc kubenswrapper[4719]: I1215 12:19:53.257744 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"97dac188-c1d2-4256-8aed-d4519e718ecd","Type":"ContainerDied","Data":"449e271bb75e932b6ce65fb72c24c9972a68d4c7b4300cf944345ed56c359173"} Dec 15 12:19:53 crc kubenswrapper[4719]: I1215 12:19:53.260176 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d649d11f-b46b-4535-a07f-99089284ace8","Type":"ContainerStarted","Data":"7b588984f97a6e783a90dac2c21017e684cc6aa6d3bbe9f6e8bf14f7a03de98e"} Dec 15 12:19:53 crc kubenswrapper[4719]: I1215 12:19:53.292558 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.292537669 podStartE2EDuration="3.292537669s" podCreationTimestamp="2025-12-15 12:19:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:19:53.289214279 +0000 UTC m=+154.231507319" watchObservedRunningTime="2025-12-15 12:19:53.292537669 +0000 UTC m=+154.234830699" Dec 15 12:19:53 crc kubenswrapper[4719]: I1215 12:19:53.631712 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:53 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:53 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:53 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:53 crc kubenswrapper[4719]: I1215 12:19:53.631776 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:54 crc kubenswrapper[4719]: I1215 12:19:54.197893 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-827sb" Dec 15 12:19:54 crc kubenswrapper[4719]: I1215 12:19:54.315737 4719 generic.go:334] "Generic (PLEG): container finished" podID="d649d11f-b46b-4535-a07f-99089284ace8" containerID="7b588984f97a6e783a90dac2c21017e684cc6aa6d3bbe9f6e8bf14f7a03de98e" exitCode=0 Dec 15 12:19:54 crc kubenswrapper[4719]: I1215 12:19:54.316037 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d649d11f-b46b-4535-a07f-99089284ace8","Type":"ContainerDied","Data":"7b588984f97a6e783a90dac2c21017e684cc6aa6d3bbe9f6e8bf14f7a03de98e"} Dec 15 12:19:54 crc kubenswrapper[4719]: I1215 12:19:54.635382 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:54 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:54 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:54 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:54 crc kubenswrapper[4719]: I1215 12:19:54.635665 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:54 crc kubenswrapper[4719]: I1215 12:19:54.900082 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 15 12:19:54 crc kubenswrapper[4719]: I1215 12:19:54.979666 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97dac188-c1d2-4256-8aed-d4519e718ecd-kubelet-dir\") pod \"97dac188-c1d2-4256-8aed-d4519e718ecd\" (UID: \"97dac188-c1d2-4256-8aed-d4519e718ecd\") " Dec 15 12:19:54 crc kubenswrapper[4719]: I1215 12:19:54.979750 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97dac188-c1d2-4256-8aed-d4519e718ecd-kube-api-access\") pod \"97dac188-c1d2-4256-8aed-d4519e718ecd\" (UID: \"97dac188-c1d2-4256-8aed-d4519e718ecd\") " Dec 15 12:19:54 crc kubenswrapper[4719]: I1215 12:19:54.980007 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97dac188-c1d2-4256-8aed-d4519e718ecd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "97dac188-c1d2-4256-8aed-d4519e718ecd" (UID: "97dac188-c1d2-4256-8aed-d4519e718ecd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:19:55 crc kubenswrapper[4719]: I1215 12:19:55.000740 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97dac188-c1d2-4256-8aed-d4519e718ecd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "97dac188-c1d2-4256-8aed-d4519e718ecd" (UID: "97dac188-c1d2-4256-8aed-d4519e718ecd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:19:55 crc kubenswrapper[4719]: I1215 12:19:55.081276 4719 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97dac188-c1d2-4256-8aed-d4519e718ecd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 15 12:19:55 crc kubenswrapper[4719]: I1215 12:19:55.081327 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97dac188-c1d2-4256-8aed-d4519e718ecd-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 15 12:19:55 crc kubenswrapper[4719]: I1215 12:19:55.323818 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"97dac188-c1d2-4256-8aed-d4519e718ecd","Type":"ContainerDied","Data":"f49d9959bf5aa5687ed3a10e093c8ef84f0272b758a0108858415935a1204a07"} Dec 15 12:19:55 crc kubenswrapper[4719]: I1215 12:19:55.323838 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 15 12:19:55 crc kubenswrapper[4719]: I1215 12:19:55.323872 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f49d9959bf5aa5687ed3a10e093c8ef84f0272b758a0108858415935a1204a07" Dec 15 12:19:55 crc kubenswrapper[4719]: I1215 12:19:55.636033 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:55 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:55 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:55 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:55 crc kubenswrapper[4719]: I1215 12:19:55.636250 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:56 crc kubenswrapper[4719]: I1215 12:19:56.635192 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:56 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:56 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:56 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:56 crc kubenswrapper[4719]: I1215 12:19:56.635257 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:57 crc kubenswrapper[4719]: I1215 12:19:57.631404 4719 patch_prober.go:28] interesting pod/router-default-5444994796-jkk5f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 15 12:19:57 crc kubenswrapper[4719]: [-]has-synced failed: reason withheld Dec 15 12:19:57 crc kubenswrapper[4719]: [+]process-running ok Dec 15 12:19:57 crc kubenswrapper[4719]: healthz check failed Dec 15 12:19:57 crc kubenswrapper[4719]: I1215 12:19:57.632147 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jkk5f" podUID="99aced59-b6b7-4abc-8d73-2dc483a14f8f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 15 12:19:58 crc kubenswrapper[4719]: I1215 12:19:58.684159 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:58 crc kubenswrapper[4719]: I1215 12:19:58.693897 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-jkk5f" Dec 15 12:19:59 crc kubenswrapper[4719]: I1215 12:19:59.059221 4719 patch_prober.go:28] interesting pod/downloads-7954f5f757-glf8c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 15 12:19:59 crc kubenswrapper[4719]: I1215 12:19:59.059552 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-glf8c" podUID="daadac3d-d689-4424-a6ff-1a6da646d03f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 15 12:19:59 crc kubenswrapper[4719]: I1215 12:19:59.059364 4719 patch_prober.go:28] interesting pod/downloads-7954f5f757-glf8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 15 12:19:59 crc kubenswrapper[4719]: I1215 12:19:59.059648 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-glf8c" podUID="daadac3d-d689-4424-a6ff-1a6da646d03f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 15 12:19:59 crc kubenswrapper[4719]: I1215 12:19:59.215234 4719 patch_prober.go:28] interesting pod/console-f9d7485db-2xxrl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 15 12:19:59 crc kubenswrapper[4719]: I1215 12:19:59.215284 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2xxrl" podUID="c8458395-1102-446b-8b75-752820a9cfdb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 15 12:20:05 crc kubenswrapper[4719]: I1215 12:20:05.994582 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:20:06 crc kubenswrapper[4719]: I1215 12:20:06.000423 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0f0b6df0-0612-46c9-83e3-b96f0ec5978d-metrics-certs\") pod \"network-metrics-daemon-9kh99\" (UID: \"0f0b6df0-0612-46c9-83e3-b96f0ec5978d\") " pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:20:06 crc kubenswrapper[4719]: I1215 12:20:06.042379 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9kh99" Dec 15 12:20:09 crc kubenswrapper[4719]: I1215 12:20:09.076072 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-glf8c" Dec 15 12:20:09 crc kubenswrapper[4719]: I1215 12:20:09.446044 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:20:09 crc kubenswrapper[4719]: I1215 12:20:09.453426 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:20:10 crc kubenswrapper[4719]: I1215 12:20:10.401145 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:20:14 crc kubenswrapper[4719]: I1215 12:20:14.228901 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 15 12:20:14 crc kubenswrapper[4719]: I1215 12:20:14.330033 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d649d11f-b46b-4535-a07f-99089284ace8-kubelet-dir\") pod \"d649d11f-b46b-4535-a07f-99089284ace8\" (UID: \"d649d11f-b46b-4535-a07f-99089284ace8\") " Dec 15 12:20:14 crc kubenswrapper[4719]: I1215 12:20:14.330352 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d649d11f-b46b-4535-a07f-99089284ace8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d649d11f-b46b-4535-a07f-99089284ace8" (UID: "d649d11f-b46b-4535-a07f-99089284ace8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:20:14 crc kubenswrapper[4719]: I1215 12:20:14.330437 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d649d11f-b46b-4535-a07f-99089284ace8-kube-api-access\") pod \"d649d11f-b46b-4535-a07f-99089284ace8\" (UID: \"d649d11f-b46b-4535-a07f-99089284ace8\") " Dec 15 12:20:14 crc kubenswrapper[4719]: I1215 12:20:14.330696 4719 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d649d11f-b46b-4535-a07f-99089284ace8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:14 crc kubenswrapper[4719]: I1215 12:20:14.335241 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d649d11f-b46b-4535-a07f-99089284ace8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d649d11f-b46b-4535-a07f-99089284ace8" (UID: "d649d11f-b46b-4535-a07f-99089284ace8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:20:14 crc kubenswrapper[4719]: I1215 12:20:14.431665 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d649d11f-b46b-4535-a07f-99089284ace8-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:14 crc kubenswrapper[4719]: I1215 12:20:14.469192 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d649d11f-b46b-4535-a07f-99089284ace8","Type":"ContainerDied","Data":"bc9fcd8bc405785b48c6721de171f97850a230494c1b67601765f67e5056eece"} Dec 15 12:20:14 crc kubenswrapper[4719]: I1215 12:20:14.469228 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc9fcd8bc405785b48c6721de171f97850a230494c1b67601765f67e5056eece" Dec 15 12:20:14 crc kubenswrapper[4719]: I1215 12:20:14.469317 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 15 12:20:19 crc kubenswrapper[4719]: I1215 12:20:19.089501 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8wqj5" Dec 15 12:20:19 crc kubenswrapper[4719]: E1215 12:20:19.973134 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:c473f3256470f7229573fb68bfdf313bc1906196d19f06a0567c03db0174c812: Get \"https://registry.redhat.io/v2/redhat/certified-operator-index/blobs/sha256:c473f3256470f7229573fb68bfdf313bc1906196d19f06a0567c03db0174c812\": context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 15 12:20:19 crc kubenswrapper[4719]: E1215 12:20:19.973742 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lzjhp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-vqndj_openshift-marketplace(722c74db-78bf-49c2-8ff7-03cba8ea3afc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:c473f3256470f7229573fb68bfdf313bc1906196d19f06a0567c03db0174c812: Get \"https://registry.redhat.io/v2/redhat/certified-operator-index/blobs/sha256:c473f3256470f7229573fb68bfdf313bc1906196d19f06a0567c03db0174c812\": context canceled" logger="UnhandledError" Dec 15 12:20:19 crc kubenswrapper[4719]: E1215 12:20:19.975029 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:c473f3256470f7229573fb68bfdf313bc1906196d19f06a0567c03db0174c812: Get \\\"https://registry.redhat.io/v2/redhat/certified-operator-index/blobs/sha256:c473f3256470f7229573fb68bfdf313bc1906196d19f06a0567c03db0174c812\\\": context canceled\"" pod="openshift-marketplace/certified-operators-vqndj" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" Dec 15 12:20:21 crc kubenswrapper[4719]: I1215 12:20:21.356826 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:20:21 crc kubenswrapper[4719]: I1215 12:20:21.357213 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:20:21 crc kubenswrapper[4719]: E1215 12:20:21.757549 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-vqndj" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" Dec 15 12:20:23 crc kubenswrapper[4719]: E1215 12:20:23.206024 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 15 12:20:23 crc kubenswrapper[4719]: E1215 12:20:23.206402 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qvnd5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-z2htt_openshift-marketplace(b166625d-5aa1-46c1-9ec7-54aec2503fb7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 15 12:20:23 crc kubenswrapper[4719]: E1215 12:20:23.207820 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-z2htt" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" Dec 15 12:20:24 crc kubenswrapper[4719]: E1215 12:20:24.202254 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-z2htt" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" Dec 15 12:20:24 crc kubenswrapper[4719]: E1215 12:20:24.264431 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 15 12:20:24 crc kubenswrapper[4719]: E1215 12:20:24.264558 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5n78w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ckh9g_openshift-marketplace(ea0b8649-9e2e-46d2-b9f0-5c51f155891e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 15 12:20:24 crc kubenswrapper[4719]: E1215 12:20:24.266125 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ckh9g" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" Dec 15 12:20:24 crc kubenswrapper[4719]: E1215 12:20:24.314232 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 15 12:20:24 crc kubenswrapper[4719]: E1215 12:20:24.314366 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xbjvv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-k2xhs_openshift-marketplace(cdd254ce-5006-40e1-a7f8-0850d049242f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 15 12:20:24 crc kubenswrapper[4719]: E1215 12:20:24.315743 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-k2xhs" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" Dec 15 12:20:26 crc kubenswrapper[4719]: I1215 12:20:26.726263 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 15 12:20:28 crc kubenswrapper[4719]: E1215 12:20:28.213510 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ckh9g" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" Dec 15 12:20:28 crc kubenswrapper[4719]: E1215 12:20:28.213801 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-k2xhs" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" Dec 15 12:20:28 crc kubenswrapper[4719]: E1215 12:20:28.266705 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 15 12:20:28 crc kubenswrapper[4719]: E1215 12:20:28.272093 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cbdq8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-kh6zf_openshift-marketplace(20547bc3-5b4d-40fd-898e-49b148705327): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 15 12:20:28 crc kubenswrapper[4719]: E1215 12:20:28.273844 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-kh6zf" podUID="20547bc3-5b4d-40fd-898e-49b148705327" Dec 15 12:20:28 crc kubenswrapper[4719]: I1215 12:20:28.553657 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r6hk" event={"ID":"f162837e-a473-4ee1-ac47-3caca3b12bbc","Type":"ContainerStarted","Data":"40ff66709abff6e6cca95c3bdecf1f2a7de542bb643633aeb5d948c7646b9cb4"} Dec 15 12:20:28 crc kubenswrapper[4719]: E1215 12:20:28.556151 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kh6zf" podUID="20547bc3-5b4d-40fd-898e-49b148705327" Dec 15 12:20:28 crc kubenswrapper[4719]: I1215 12:20:28.635353 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9kh99"] Dec 15 12:20:28 crc kubenswrapper[4719]: W1215 12:20:28.641395 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f0b6df0_0612_46c9_83e3_b96f0ec5978d.slice/crio-9e6e5cc8256e920693c87542a1e4c9ed74cc6714945ca52b58ca2060b5bb1d53 WatchSource:0}: Error finding container 9e6e5cc8256e920693c87542a1e4c9ed74cc6714945ca52b58ca2060b5bb1d53: Status 404 returned error can't find the container with id 9e6e5cc8256e920693c87542a1e4c9ed74cc6714945ca52b58ca2060b5bb1d53 Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.559049 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9kh99" event={"ID":"0f0b6df0-0612-46c9-83e3-b96f0ec5978d","Type":"ContainerStarted","Data":"9e6e5cc8256e920693c87542a1e4c9ed74cc6714945ca52b58ca2060b5bb1d53"} Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.561214 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmfbz" event={"ID":"6014e2e6-c315-4f83-9266-33239e0c43c4","Type":"ContainerStarted","Data":"52338cdac04218ac5d986eae0c2e60c8159a057403d68b3dee09f6e138b8fffa"} Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.563113 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gwjsp" event={"ID":"638821b0-deb7-40ee-bad1-b552994b4598","Type":"ContainerStarted","Data":"4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53"} Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.638755 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 15 12:20:29 crc kubenswrapper[4719]: E1215 12:20:29.639081 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d649d11f-b46b-4535-a07f-99089284ace8" containerName="pruner" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.639105 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d649d11f-b46b-4535-a07f-99089284ace8" containerName="pruner" Dec 15 12:20:29 crc kubenswrapper[4719]: E1215 12:20:29.639127 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b7161e6-9af7-4443-9af7-2d11e0f36d68" containerName="collect-profiles" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.639141 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b7161e6-9af7-4443-9af7-2d11e0f36d68" containerName="collect-profiles" Dec 15 12:20:29 crc kubenswrapper[4719]: E1215 12:20:29.639162 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97dac188-c1d2-4256-8aed-d4519e718ecd" containerName="pruner" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.639173 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="97dac188-c1d2-4256-8aed-d4519e718ecd" containerName="pruner" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.639333 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="97dac188-c1d2-4256-8aed-d4519e718ecd" containerName="pruner" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.639350 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d649d11f-b46b-4535-a07f-99089284ace8" containerName="pruner" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.639369 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b7161e6-9af7-4443-9af7-2d11e0f36d68" containerName="collect-profiles" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.640753 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.648724 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.649278 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.649571 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.738593 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc705873-1a38-44cf-bab8-75971af8a392-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc705873-1a38-44cf-bab8-75971af8a392\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.738660 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc705873-1a38-44cf-bab8-75971af8a392-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc705873-1a38-44cf-bab8-75971af8a392\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.839293 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc705873-1a38-44cf-bab8-75971af8a392-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc705873-1a38-44cf-bab8-75971af8a392\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.839362 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc705873-1a38-44cf-bab8-75971af8a392-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc705873-1a38-44cf-bab8-75971af8a392\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.839457 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc705873-1a38-44cf-bab8-75971af8a392-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc705873-1a38-44cf-bab8-75971af8a392\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.856838 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc705873-1a38-44cf-bab8-75971af8a392-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc705873-1a38-44cf-bab8-75971af8a392\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 15 12:20:29 crc kubenswrapper[4719]: I1215 12:20:29.974083 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 15 12:20:30 crc kubenswrapper[4719]: I1215 12:20:30.569529 4719 generic.go:334] "Generic (PLEG): container finished" podID="638821b0-deb7-40ee-bad1-b552994b4598" containerID="4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53" exitCode=0 Dec 15 12:20:30 crc kubenswrapper[4719]: I1215 12:20:30.569876 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gwjsp" event={"ID":"638821b0-deb7-40ee-bad1-b552994b4598","Type":"ContainerDied","Data":"4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53"} Dec 15 12:20:30 crc kubenswrapper[4719]: I1215 12:20:30.574187 4719 generic.go:334] "Generic (PLEG): container finished" podID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerID="40ff66709abff6e6cca95c3bdecf1f2a7de542bb643633aeb5d948c7646b9cb4" exitCode=0 Dec 15 12:20:30 crc kubenswrapper[4719]: I1215 12:20:30.574238 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r6hk" event={"ID":"f162837e-a473-4ee1-ac47-3caca3b12bbc","Type":"ContainerDied","Data":"40ff66709abff6e6cca95c3bdecf1f2a7de542bb643633aeb5d948c7646b9cb4"} Dec 15 12:20:30 crc kubenswrapper[4719]: I1215 12:20:30.576870 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9kh99" event={"ID":"0f0b6df0-0612-46c9-83e3-b96f0ec5978d","Type":"ContainerStarted","Data":"c4f1ea08316eb3817130a0511e9e5552d8762d6e8c1a9d187a162d7b0e2e1181"} Dec 15 12:20:30 crc kubenswrapper[4719]: I1215 12:20:30.576895 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9kh99" event={"ID":"0f0b6df0-0612-46c9-83e3-b96f0ec5978d","Type":"ContainerStarted","Data":"ae44caada9685071f15df74df08ce4a7af79220e593c001d4ed0d71fae53bc19"} Dec 15 12:20:30 crc kubenswrapper[4719]: I1215 12:20:30.578617 4719 generic.go:334] "Generic (PLEG): container finished" podID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerID="52338cdac04218ac5d986eae0c2e60c8159a057403d68b3dee09f6e138b8fffa" exitCode=0 Dec 15 12:20:30 crc kubenswrapper[4719]: I1215 12:20:30.578641 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmfbz" event={"ID":"6014e2e6-c315-4f83-9266-33239e0c43c4","Type":"ContainerDied","Data":"52338cdac04218ac5d986eae0c2e60c8159a057403d68b3dee09f6e138b8fffa"} Dec 15 12:20:31 crc kubenswrapper[4719]: I1215 12:20:31.207030 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 15 12:20:31 crc kubenswrapper[4719]: I1215 12:20:31.589296 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bc705873-1a38-44cf-bab8-75971af8a392","Type":"ContainerStarted","Data":"b650c7347f4ace6a05ca1d4323fa79a91e646c8d7cf41e4f8584b9bec9435531"} Dec 15 12:20:31 crc kubenswrapper[4719]: I1215 12:20:31.608748 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-9kh99" podStartSLOduration=168.608733871 podStartE2EDuration="2m48.608733871s" podCreationTimestamp="2025-12-15 12:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:20:31.605963969 +0000 UTC m=+192.548256999" watchObservedRunningTime="2025-12-15 12:20:31.608733871 +0000 UTC m=+192.551026901" Dec 15 12:20:32 crc kubenswrapper[4719]: I1215 12:20:32.598110 4719 generic.go:334] "Generic (PLEG): container finished" podID="bc705873-1a38-44cf-bab8-75971af8a392" containerID="1bb18e8f78e7c0606ce60a38b6dc50c314e63d99afd234386cfef70021aca3a7" exitCode=0 Dec 15 12:20:32 crc kubenswrapper[4719]: I1215 12:20:32.598228 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bc705873-1a38-44cf-bab8-75971af8a392","Type":"ContainerDied","Data":"1bb18e8f78e7c0606ce60a38b6dc50c314e63d99afd234386cfef70021aca3a7"} Dec 15 12:20:33 crc kubenswrapper[4719]: I1215 12:20:33.606582 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r6hk" event={"ID":"f162837e-a473-4ee1-ac47-3caca3b12bbc","Type":"ContainerStarted","Data":"971834e51d361567c2809cadb70fc753f58d9eca59d4d12f4f8d5c4b9c7dda03"} Dec 15 12:20:33 crc kubenswrapper[4719]: I1215 12:20:33.886062 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 15 12:20:33 crc kubenswrapper[4719]: I1215 12:20:33.899036 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5r6hk" podStartSLOduration=5.556565313 podStartE2EDuration="46.899019617s" podCreationTimestamp="2025-12-15 12:19:47 +0000 UTC" firstStartedPulling="2025-12-15 12:19:51.136986429 +0000 UTC m=+152.079279459" lastFinishedPulling="2025-12-15 12:20:32.479440733 +0000 UTC m=+193.421733763" observedRunningTime="2025-12-15 12:20:33.624040323 +0000 UTC m=+194.566333363" watchObservedRunningTime="2025-12-15 12:20:33.899019617 +0000 UTC m=+194.841312647" Dec 15 12:20:33 crc kubenswrapper[4719]: I1215 12:20:33.993428 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc705873-1a38-44cf-bab8-75971af8a392-kube-api-access\") pod \"bc705873-1a38-44cf-bab8-75971af8a392\" (UID: \"bc705873-1a38-44cf-bab8-75971af8a392\") " Dec 15 12:20:33 crc kubenswrapper[4719]: I1215 12:20:33.993507 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc705873-1a38-44cf-bab8-75971af8a392-kubelet-dir\") pod \"bc705873-1a38-44cf-bab8-75971af8a392\" (UID: \"bc705873-1a38-44cf-bab8-75971af8a392\") " Dec 15 12:20:33 crc kubenswrapper[4719]: I1215 12:20:33.993624 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc705873-1a38-44cf-bab8-75971af8a392-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bc705873-1a38-44cf-bab8-75971af8a392" (UID: "bc705873-1a38-44cf-bab8-75971af8a392"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:20:33 crc kubenswrapper[4719]: I1215 12:20:33.993866 4719 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc705873-1a38-44cf-bab8-75971af8a392-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:33 crc kubenswrapper[4719]: I1215 12:20:33.999201 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc705873-1a38-44cf-bab8-75971af8a392-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bc705873-1a38-44cf-bab8-75971af8a392" (UID: "bc705873-1a38-44cf-bab8-75971af8a392"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.095300 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc705873-1a38-44cf-bab8-75971af8a392-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.436108 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 15 12:20:34 crc kubenswrapper[4719]: E1215 12:20:34.436377 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc705873-1a38-44cf-bab8-75971af8a392" containerName="pruner" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.436401 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc705873-1a38-44cf-bab8-75971af8a392" containerName="pruner" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.436809 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc705873-1a38-44cf-bab8-75971af8a392" containerName="pruner" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.440844 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.449096 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.500030 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6426eb80-427c-411c-bf47-af018703e000-kube-api-access\") pod \"installer-9-crc\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.500149 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-var-lock\") pod \"installer-9-crc\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.500173 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.601230 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-var-lock\") pod \"installer-9-crc\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.601270 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.601308 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6426eb80-427c-411c-bf47-af018703e000-kube-api-access\") pod \"installer-9-crc\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.601601 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-var-lock\") pod \"installer-9-crc\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.601639 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.612301 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bc705873-1a38-44cf-bab8-75971af8a392","Type":"ContainerDied","Data":"b650c7347f4ace6a05ca1d4323fa79a91e646c8d7cf41e4f8584b9bec9435531"} Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.612335 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.612342 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b650c7347f4ace6a05ca1d4323fa79a91e646c8d7cf41e4f8584b9bec9435531" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.621757 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6426eb80-427c-411c-bf47-af018703e000-kube-api-access\") pod \"installer-9-crc\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:34 crc kubenswrapper[4719]: I1215 12:20:34.764278 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:20:35 crc kubenswrapper[4719]: I1215 12:20:35.158044 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 15 12:20:35 crc kubenswrapper[4719]: I1215 12:20:35.618566 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6426eb80-427c-411c-bf47-af018703e000","Type":"ContainerStarted","Data":"eb61cca2433ad63dc066466de67e24311832bf1e0963c9ed0e85648cc7375fd4"} Dec 15 12:20:36 crc kubenswrapper[4719]: I1215 12:20:36.626971 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmfbz" event={"ID":"6014e2e6-c315-4f83-9266-33239e0c43c4","Type":"ContainerStarted","Data":"e42ce08bbabdc3c0d285a2edd5768510690480cfda6d0a2defaf69b3f0f7eefc"} Dec 15 12:20:36 crc kubenswrapper[4719]: I1215 12:20:36.647898 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tmfbz" podStartSLOduration=4.0187033 podStartE2EDuration="46.647876124s" podCreationTimestamp="2025-12-15 12:19:50 +0000 UTC" firstStartedPulling="2025-12-15 12:19:52.237579657 +0000 UTC m=+153.179872687" lastFinishedPulling="2025-12-15 12:20:34.866752471 +0000 UTC m=+195.809045511" observedRunningTime="2025-12-15 12:20:36.644320366 +0000 UTC m=+197.586613416" watchObservedRunningTime="2025-12-15 12:20:36.647876124 +0000 UTC m=+197.590169164" Dec 15 12:20:38 crc kubenswrapper[4719]: I1215 12:20:38.505155 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:20:38 crc kubenswrapper[4719]: I1215 12:20:38.505500 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:20:38 crc kubenswrapper[4719]: I1215 12:20:38.616485 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:20:38 crc kubenswrapper[4719]: I1215 12:20:38.638548 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gwjsp" event={"ID":"638821b0-deb7-40ee-bad1-b552994b4598","Type":"ContainerStarted","Data":"8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f"} Dec 15 12:20:38 crc kubenswrapper[4719]: I1215 12:20:38.640179 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6426eb80-427c-411c-bf47-af018703e000","Type":"ContainerStarted","Data":"0c58d556efb0decf61335e8d184efe3c5da120463bab307bab767c3acf82c75a"} Dec 15 12:20:38 crc kubenswrapper[4719]: I1215 12:20:38.642611 4719 generic.go:334] "Generic (PLEG): container finished" podID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerID="0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d" exitCode=0 Dec 15 12:20:38 crc kubenswrapper[4719]: I1215 12:20:38.643097 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2htt" event={"ID":"b166625d-5aa1-46c1-9ec7-54aec2503fb7","Type":"ContainerDied","Data":"0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d"} Dec 15 12:20:38 crc kubenswrapper[4719]: I1215 12:20:38.682060 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=4.6820421119999995 podStartE2EDuration="4.682042112s" podCreationTimestamp="2025-12-15 12:20:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:20:38.681413822 +0000 UTC m=+199.623706852" watchObservedRunningTime="2025-12-15 12:20:38.682042112 +0000 UTC m=+199.624335142" Dec 15 12:20:38 crc kubenswrapper[4719]: I1215 12:20:38.683879 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gwjsp" podStartSLOduration=4.86425904 podStartE2EDuration="49.683850533s" podCreationTimestamp="2025-12-15 12:19:49 +0000 UTC" firstStartedPulling="2025-12-15 12:19:52.232119816 +0000 UTC m=+153.174412846" lastFinishedPulling="2025-12-15 12:20:37.051711309 +0000 UTC m=+197.994004339" observedRunningTime="2025-12-15 12:20:38.660815017 +0000 UTC m=+199.603108047" watchObservedRunningTime="2025-12-15 12:20:38.683850533 +0000 UTC m=+199.626143573" Dec 15 12:20:38 crc kubenswrapper[4719]: I1215 12:20:38.688871 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:20:39 crc kubenswrapper[4719]: I1215 12:20:39.192134 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5r6hk"] Dec 15 12:20:39 crc kubenswrapper[4719]: I1215 12:20:39.649516 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2htt" event={"ID":"b166625d-5aa1-46c1-9ec7-54aec2503fb7","Type":"ContainerStarted","Data":"bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4"} Dec 15 12:20:39 crc kubenswrapper[4719]: I1215 12:20:39.651438 4719 generic.go:334] "Generic (PLEG): container finished" podID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerID="346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94" exitCode=0 Dec 15 12:20:39 crc kubenswrapper[4719]: I1215 12:20:39.651519 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqndj" event={"ID":"722c74db-78bf-49c2-8ff7-03cba8ea3afc","Type":"ContainerDied","Data":"346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94"} Dec 15 12:20:39 crc kubenswrapper[4719]: I1215 12:20:39.675039 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z2htt" podStartSLOduration=4.55456363 podStartE2EDuration="52.675023086s" podCreationTimestamp="2025-12-15 12:19:47 +0000 UTC" firstStartedPulling="2025-12-15 12:19:51.143376192 +0000 UTC m=+152.085669222" lastFinishedPulling="2025-12-15 12:20:39.263835658 +0000 UTC m=+200.206128678" observedRunningTime="2025-12-15 12:20:39.672846183 +0000 UTC m=+200.615139213" watchObservedRunningTime="2025-12-15 12:20:39.675023086 +0000 UTC m=+200.617316116" Dec 15 12:20:40 crc kubenswrapper[4719]: I1215 12:20:40.457222 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:20:40 crc kubenswrapper[4719]: I1215 12:20:40.457532 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:20:40 crc kubenswrapper[4719]: I1215 12:20:40.582486 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:20:40 crc kubenswrapper[4719]: I1215 12:20:40.582521 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:20:40 crc kubenswrapper[4719]: I1215 12:20:40.660190 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqndj" event={"ID":"722c74db-78bf-49c2-8ff7-03cba8ea3afc","Type":"ContainerStarted","Data":"5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594"} Dec 15 12:20:40 crc kubenswrapper[4719]: I1215 12:20:40.662521 4719 generic.go:334] "Generic (PLEG): container finished" podID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerID="a17e8a1a7d1cb41f1ecc5d7a042a2eb37a3ded082f828bd56939b1eae4f6884d" exitCode=0 Dec 15 12:20:40 crc kubenswrapper[4719]: I1215 12:20:40.662600 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckh9g" event={"ID":"ea0b8649-9e2e-46d2-b9f0-5c51f155891e","Type":"ContainerDied","Data":"a17e8a1a7d1cb41f1ecc5d7a042a2eb37a3ded082f828bd56939b1eae4f6884d"} Dec 15 12:20:40 crc kubenswrapper[4719]: I1215 12:20:40.662742 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5r6hk" podUID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerName="registry-server" containerID="cri-o://971834e51d361567c2809cadb70fc753f58d9eca59d4d12f4f8d5c4b9c7dda03" gracePeriod=2 Dec 15 12:20:40 crc kubenswrapper[4719]: I1215 12:20:40.690901 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vqndj" podStartSLOduration=4.764209154 podStartE2EDuration="53.690623271s" podCreationTimestamp="2025-12-15 12:19:47 +0000 UTC" firstStartedPulling="2025-12-15 12:19:51.13671143 +0000 UTC m=+152.079004470" lastFinishedPulling="2025-12-15 12:20:40.063125557 +0000 UTC m=+201.005418587" observedRunningTime="2025-12-15 12:20:40.686721631 +0000 UTC m=+201.629014661" watchObservedRunningTime="2025-12-15 12:20:40.690623271 +0000 UTC m=+201.632916321" Dec 15 12:20:41 crc kubenswrapper[4719]: I1215 12:20:41.496013 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gwjsp" podUID="638821b0-deb7-40ee-bad1-b552994b4598" containerName="registry-server" probeResult="failure" output=< Dec 15 12:20:41 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 12:20:41 crc kubenswrapper[4719]: > Dec 15 12:20:41 crc kubenswrapper[4719]: I1215 12:20:41.630562 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tmfbz" podUID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerName="registry-server" probeResult="failure" output=< Dec 15 12:20:41 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 12:20:41 crc kubenswrapper[4719]: > Dec 15 12:20:42 crc kubenswrapper[4719]: I1215 12:20:42.679134 4719 generic.go:334] "Generic (PLEG): container finished" podID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerID="971834e51d361567c2809cadb70fc753f58d9eca59d4d12f4f8d5c4b9c7dda03" exitCode=0 Dec 15 12:20:42 crc kubenswrapper[4719]: I1215 12:20:42.679513 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r6hk" event={"ID":"f162837e-a473-4ee1-ac47-3caca3b12bbc","Type":"ContainerDied","Data":"971834e51d361567c2809cadb70fc753f58d9eca59d4d12f4f8d5c4b9c7dda03"} Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.086207 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.130547 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-catalog-content\") pod \"f162837e-a473-4ee1-ac47-3caca3b12bbc\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.130609 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-utilities\") pod \"f162837e-a473-4ee1-ac47-3caca3b12bbc\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.130675 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25lx5\" (UniqueName: \"kubernetes.io/projected/f162837e-a473-4ee1-ac47-3caca3b12bbc-kube-api-access-25lx5\") pod \"f162837e-a473-4ee1-ac47-3caca3b12bbc\" (UID: \"f162837e-a473-4ee1-ac47-3caca3b12bbc\") " Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.131472 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-utilities" (OuterVolumeSpecName: "utilities") pod "f162837e-a473-4ee1-ac47-3caca3b12bbc" (UID: "f162837e-a473-4ee1-ac47-3caca3b12bbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.152095 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f162837e-a473-4ee1-ac47-3caca3b12bbc-kube-api-access-25lx5" (OuterVolumeSpecName: "kube-api-access-25lx5") pod "f162837e-a473-4ee1-ac47-3caca3b12bbc" (UID: "f162837e-a473-4ee1-ac47-3caca3b12bbc"). InnerVolumeSpecName "kube-api-access-25lx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.196184 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f162837e-a473-4ee1-ac47-3caca3b12bbc" (UID: "f162837e-a473-4ee1-ac47-3caca3b12bbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.231492 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.231521 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f162837e-a473-4ee1-ac47-3caca3b12bbc-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.231532 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25lx5\" (UniqueName: \"kubernetes.io/projected/f162837e-a473-4ee1-ac47-3caca3b12bbc-kube-api-access-25lx5\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.687551 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckh9g" event={"ID":"ea0b8649-9e2e-46d2-b9f0-5c51f155891e","Type":"ContainerStarted","Data":"4aebf3e95a84f4e4994731b6b66eda0f310be5c0cea52f1f6a5ae595eb11476c"} Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.692215 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r6hk" event={"ID":"f162837e-a473-4ee1-ac47-3caca3b12bbc","Type":"ContainerDied","Data":"88382695961da997124f15894e15ffcc84b891e37e8669db3e4e2f23c45366d9"} Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.692246 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r6hk" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.692275 4719 scope.go:117] "RemoveContainer" containerID="971834e51d361567c2809cadb70fc753f58d9eca59d4d12f4f8d5c4b9c7dda03" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.712334 4719 scope.go:117] "RemoveContainer" containerID="40ff66709abff6e6cca95c3bdecf1f2a7de542bb643633aeb5d948c7646b9cb4" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.712329 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ckh9g" podStartSLOduration=3.03441781 podStartE2EDuration="54.712308101s" podCreationTimestamp="2025-12-15 12:19:49 +0000 UTC" firstStartedPulling="2025-12-15 12:19:51.127695281 +0000 UTC m=+152.069988311" lastFinishedPulling="2025-12-15 12:20:42.805585572 +0000 UTC m=+203.747878602" observedRunningTime="2025-12-15 12:20:43.709347523 +0000 UTC m=+204.651640563" watchObservedRunningTime="2025-12-15 12:20:43.712308101 +0000 UTC m=+204.654601131" Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.724577 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5r6hk"] Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.739345 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5r6hk"] Dec 15 12:20:43 crc kubenswrapper[4719]: I1215 12:20:43.753707 4719 scope.go:117] "RemoveContainer" containerID="188aa5738cf207b7e96ceab78fbf7778d4a8d1d8b50d71248a6eb7b01a014d9e" Dec 15 12:20:44 crc kubenswrapper[4719]: I1215 12:20:44.703325 4719 generic.go:334] "Generic (PLEG): container finished" podID="20547bc3-5b4d-40fd-898e-49b148705327" containerID="cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546" exitCode=0 Dec 15 12:20:44 crc kubenswrapper[4719]: I1215 12:20:44.703982 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kh6zf" event={"ID":"20547bc3-5b4d-40fd-898e-49b148705327","Type":"ContainerDied","Data":"cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546"} Dec 15 12:20:45 crc kubenswrapper[4719]: I1215 12:20:45.484150 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f162837e-a473-4ee1-ac47-3caca3b12bbc" path="/var/lib/kubelet/pods/f162837e-a473-4ee1-ac47-3caca3b12bbc/volumes" Dec 15 12:20:48 crc kubenswrapper[4719]: I1215 12:20:48.360894 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:20:48 crc kubenswrapper[4719]: I1215 12:20:48.361577 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:20:48 crc kubenswrapper[4719]: I1215 12:20:48.411246 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:20:48 crc kubenswrapper[4719]: I1215 12:20:48.427573 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:20:48 crc kubenswrapper[4719]: I1215 12:20:48.427609 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:20:48 crc kubenswrapper[4719]: I1215 12:20:48.468484 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:20:48 crc kubenswrapper[4719]: I1215 12:20:48.730073 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2xhs" event={"ID":"cdd254ce-5006-40e1-a7f8-0850d049242f","Type":"ContainerStarted","Data":"00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d"} Dec 15 12:20:48 crc kubenswrapper[4719]: I1215 12:20:48.732367 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kh6zf" event={"ID":"20547bc3-5b4d-40fd-898e-49b148705327","Type":"ContainerStarted","Data":"2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877"} Dec 15 12:20:48 crc kubenswrapper[4719]: I1215 12:20:48.777417 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:20:48 crc kubenswrapper[4719]: I1215 12:20:48.788247 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:20:49 crc kubenswrapper[4719]: I1215 12:20:49.750116 4719 generic.go:334] "Generic (PLEG): container finished" podID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerID="00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d" exitCode=0 Dec 15 12:20:49 crc kubenswrapper[4719]: I1215 12:20:49.750329 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2xhs" event={"ID":"cdd254ce-5006-40e1-a7f8-0850d049242f","Type":"ContainerDied","Data":"00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d"} Dec 15 12:20:49 crc kubenswrapper[4719]: I1215 12:20:49.771786 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kh6zf" podStartSLOduration=4.08179459 podStartE2EDuration="1m0.771771364s" podCreationTimestamp="2025-12-15 12:19:49 +0000 UTC" firstStartedPulling="2025-12-15 12:19:51.160916675 +0000 UTC m=+152.103209705" lastFinishedPulling="2025-12-15 12:20:47.850893449 +0000 UTC m=+208.793186479" observedRunningTime="2025-12-15 12:20:49.771035279 +0000 UTC m=+210.713328309" watchObservedRunningTime="2025-12-15 12:20:49.771771364 +0000 UTC m=+210.714064394" Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.041537 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.041902 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.090794 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.525734 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.595918 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.655958 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.706052 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.757216 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2xhs" event={"ID":"cdd254ce-5006-40e1-a7f8-0850d049242f","Type":"ContainerStarted","Data":"abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960"} Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.802372 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.822578 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k2xhs" podStartSLOduration=4.279185026 podStartE2EDuration="1m4.822564118s" podCreationTimestamp="2025-12-15 12:19:46 +0000 UTC" firstStartedPulling="2025-12-15 12:19:50.041944656 +0000 UTC m=+150.984237686" lastFinishedPulling="2025-12-15 12:20:50.585323748 +0000 UTC m=+211.527616778" observedRunningTime="2025-12-15 12:20:50.77357254 +0000 UTC m=+211.715865590" watchObservedRunningTime="2025-12-15 12:20:50.822564118 +0000 UTC m=+211.764857148" Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.997355 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z2htt"] Dec 15 12:20:50 crc kubenswrapper[4719]: I1215 12:20:50.997791 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z2htt" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerName="registry-server" containerID="cri-o://bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4" gracePeriod=2 Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.356438 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.357261 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.357302 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.357341 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.357821 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.357941 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9" gracePeriod=600 Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.444622 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvnd5\" (UniqueName: \"kubernetes.io/projected/b166625d-5aa1-46c1-9ec7-54aec2503fb7-kube-api-access-qvnd5\") pod \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.444718 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-utilities\") pod \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.444748 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-catalog-content\") pod \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\" (UID: \"b166625d-5aa1-46c1-9ec7-54aec2503fb7\") " Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.445702 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-utilities" (OuterVolumeSpecName: "utilities") pod "b166625d-5aa1-46c1-9ec7-54aec2503fb7" (UID: "b166625d-5aa1-46c1-9ec7-54aec2503fb7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.458011 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b166625d-5aa1-46c1-9ec7-54aec2503fb7-kube-api-access-qvnd5" (OuterVolumeSpecName: "kube-api-access-qvnd5") pod "b166625d-5aa1-46c1-9ec7-54aec2503fb7" (UID: "b166625d-5aa1-46c1-9ec7-54aec2503fb7"). InnerVolumeSpecName "kube-api-access-qvnd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.503022 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b166625d-5aa1-46c1-9ec7-54aec2503fb7" (UID: "b166625d-5aa1-46c1-9ec7-54aec2503fb7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.546065 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.546104 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b166625d-5aa1-46c1-9ec7-54aec2503fb7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.546120 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvnd5\" (UniqueName: \"kubernetes.io/projected/b166625d-5aa1-46c1-9ec7-54aec2503fb7-kube-api-access-qvnd5\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.764694 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9" exitCode=0 Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.764953 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9"} Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.764979 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"38e050f4b3eb5ec1de6c0f4f194704b5cba328b77e0bccea20d90d6077185480"} Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.767678 4719 generic.go:334] "Generic (PLEG): container finished" podID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerID="bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4" exitCode=0 Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.768134 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z2htt" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.772963 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2htt" event={"ID":"b166625d-5aa1-46c1-9ec7-54aec2503fb7","Type":"ContainerDied","Data":"bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4"} Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.773064 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2htt" event={"ID":"b166625d-5aa1-46c1-9ec7-54aec2503fb7","Type":"ContainerDied","Data":"df6eeffeb3558d26d92511ae98595ff563e4551c9540e4095e61fe1ce84ef017"} Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.773092 4719 scope.go:117] "RemoveContainer" containerID="bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.792007 4719 scope.go:117] "RemoveContainer" containerID="0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.802269 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z2htt"] Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.805215 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z2htt"] Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.812664 4719 scope.go:117] "RemoveContainer" containerID="2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.829699 4719 scope.go:117] "RemoveContainer" containerID="bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4" Dec 15 12:20:51 crc kubenswrapper[4719]: E1215 12:20:51.830308 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4\": container with ID starting with bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4 not found: ID does not exist" containerID="bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.830357 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4"} err="failed to get container status \"bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4\": rpc error: code = NotFound desc = could not find container \"bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4\": container with ID starting with bbebf49f7f58149e7603ede04949868a1dbcd1ee5af36e7025005d7c2fb2a0e4 not found: ID does not exist" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.830379 4719 scope.go:117] "RemoveContainer" containerID="0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d" Dec 15 12:20:51 crc kubenswrapper[4719]: E1215 12:20:51.831325 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d\": container with ID starting with 0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d not found: ID does not exist" containerID="0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.831348 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d"} err="failed to get container status \"0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d\": rpc error: code = NotFound desc = could not find container \"0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d\": container with ID starting with 0c4298e90a4243fdb125752ff07d60cecf2dc5b39d0eaa0d7a8bc8bb2c75783d not found: ID does not exist" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.831362 4719 scope.go:117] "RemoveContainer" containerID="2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19" Dec 15 12:20:51 crc kubenswrapper[4719]: E1215 12:20:51.831686 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19\": container with ID starting with 2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19 not found: ID does not exist" containerID="2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19" Dec 15 12:20:51 crc kubenswrapper[4719]: I1215 12:20:51.831727 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19"} err="failed to get container status \"2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19\": rpc error: code = NotFound desc = could not find container \"2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19\": container with ID starting with 2f05e561aebb6090b20a78863ea46ae24744b172cd4bdcf4ad33f4c7f3fc2f19 not found: ID does not exist" Dec 15 12:20:53 crc kubenswrapper[4719]: I1215 12:20:53.386641 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckh9g"] Dec 15 12:20:53 crc kubenswrapper[4719]: I1215 12:20:53.483186 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" path="/var/lib/kubelet/pods/b166625d-5aa1-46c1-9ec7-54aec2503fb7/volumes" Dec 15 12:20:53 crc kubenswrapper[4719]: I1215 12:20:53.589423 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tmfbz"] Dec 15 12:20:53 crc kubenswrapper[4719]: I1215 12:20:53.590081 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tmfbz" podUID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerName="registry-server" containerID="cri-o://e42ce08bbabdc3c0d285a2edd5768510690480cfda6d0a2defaf69b3f0f7eefc" gracePeriod=2 Dec 15 12:20:53 crc kubenswrapper[4719]: I1215 12:20:53.783146 4719 generic.go:334] "Generic (PLEG): container finished" podID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerID="e42ce08bbabdc3c0d285a2edd5768510690480cfda6d0a2defaf69b3f0f7eefc" exitCode=0 Dec 15 12:20:53 crc kubenswrapper[4719]: I1215 12:20:53.783205 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmfbz" event={"ID":"6014e2e6-c315-4f83-9266-33239e0c43c4","Type":"ContainerDied","Data":"e42ce08bbabdc3c0d285a2edd5768510690480cfda6d0a2defaf69b3f0f7eefc"} Dec 15 12:20:53 crc kubenswrapper[4719]: I1215 12:20:53.783370 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ckh9g" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerName="registry-server" containerID="cri-o://4aebf3e95a84f4e4994731b6b66eda0f310be5c0cea52f1f6a5ae595eb11476c" gracePeriod=2 Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.017189 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.077458 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-utilities\") pod \"6014e2e6-c315-4f83-9266-33239e0c43c4\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.077531 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hcz6\" (UniqueName: \"kubernetes.io/projected/6014e2e6-c315-4f83-9266-33239e0c43c4-kube-api-access-7hcz6\") pod \"6014e2e6-c315-4f83-9266-33239e0c43c4\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.077578 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-catalog-content\") pod \"6014e2e6-c315-4f83-9266-33239e0c43c4\" (UID: \"6014e2e6-c315-4f83-9266-33239e0c43c4\") " Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.078447 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-utilities" (OuterVolumeSpecName: "utilities") pod "6014e2e6-c315-4f83-9266-33239e0c43c4" (UID: "6014e2e6-c315-4f83-9266-33239e0c43c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.082539 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6014e2e6-c315-4f83-9266-33239e0c43c4-kube-api-access-7hcz6" (OuterVolumeSpecName: "kube-api-access-7hcz6") pod "6014e2e6-c315-4f83-9266-33239e0c43c4" (UID: "6014e2e6-c315-4f83-9266-33239e0c43c4"). InnerVolumeSpecName "kube-api-access-7hcz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.179771 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.179975 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hcz6\" (UniqueName: \"kubernetes.io/projected/6014e2e6-c315-4f83-9266-33239e0c43c4-kube-api-access-7hcz6\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.226997 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6014e2e6-c315-4f83-9266-33239e0c43c4" (UID: "6014e2e6-c315-4f83-9266-33239e0c43c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.281022 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6014e2e6-c315-4f83-9266-33239e0c43c4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.806669 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmfbz" event={"ID":"6014e2e6-c315-4f83-9266-33239e0c43c4","Type":"ContainerDied","Data":"20f7f70b96379ba9eac00aed1edfa82d1466e92f01448abfb3940efc75e5b844"} Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.806747 4719 scope.go:117] "RemoveContainer" containerID="e42ce08bbabdc3c0d285a2edd5768510690480cfda6d0a2defaf69b3f0f7eefc" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.806784 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tmfbz" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.810661 4719 generic.go:334] "Generic (PLEG): container finished" podID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerID="4aebf3e95a84f4e4994731b6b66eda0f310be5c0cea52f1f6a5ae595eb11476c" exitCode=0 Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.810709 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckh9g" event={"ID":"ea0b8649-9e2e-46d2-b9f0-5c51f155891e","Type":"ContainerDied","Data":"4aebf3e95a84f4e4994731b6b66eda0f310be5c0cea52f1f6a5ae595eb11476c"} Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.842065 4719 scope.go:117] "RemoveContainer" containerID="52338cdac04218ac5d986eae0c2e60c8159a057403d68b3dee09f6e138b8fffa" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.853603 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tmfbz"] Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.863392 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tmfbz"] Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.886146 4719 scope.go:117] "RemoveContainer" containerID="d1f3956b0b9cb7e360724843365a9a58d4a7f80f700bed717185698b68a8fcc9" Dec 15 12:20:54 crc kubenswrapper[4719]: I1215 12:20:54.984923 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.091259 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-utilities\") pod \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.091300 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-catalog-content\") pod \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.091413 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n78w\" (UniqueName: \"kubernetes.io/projected/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-kube-api-access-5n78w\") pod \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\" (UID: \"ea0b8649-9e2e-46d2-b9f0-5c51f155891e\") " Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.092731 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-utilities" (OuterVolumeSpecName: "utilities") pod "ea0b8649-9e2e-46d2-b9f0-5c51f155891e" (UID: "ea0b8649-9e2e-46d2-b9f0-5c51f155891e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.095088 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-kube-api-access-5n78w" (OuterVolumeSpecName: "kube-api-access-5n78w") pod "ea0b8649-9e2e-46d2-b9f0-5c51f155891e" (UID: "ea0b8649-9e2e-46d2-b9f0-5c51f155891e"). InnerVolumeSpecName "kube-api-access-5n78w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.108595 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea0b8649-9e2e-46d2-b9f0-5c51f155891e" (UID: "ea0b8649-9e2e-46d2-b9f0-5c51f155891e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.192930 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n78w\" (UniqueName: \"kubernetes.io/projected/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-kube-api-access-5n78w\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.192983 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.193004 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea0b8649-9e2e-46d2-b9f0-5c51f155891e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.484510 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6014e2e6-c315-4f83-9266-33239e0c43c4" path="/var/lib/kubelet/pods/6014e2e6-c315-4f83-9266-33239e0c43c4/volumes" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.819970 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckh9g" event={"ID":"ea0b8649-9e2e-46d2-b9f0-5c51f155891e","Type":"ContainerDied","Data":"34b313a0f1617e695decdbcf5461e411a96ff3013285407be7cc6b8664e2a5c6"} Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.820023 4719 scope.go:117] "RemoveContainer" containerID="4aebf3e95a84f4e4994731b6b66eda0f310be5c0cea52f1f6a5ae595eb11476c" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.820075 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckh9g" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.844057 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckh9g"] Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.845263 4719 scope.go:117] "RemoveContainer" containerID="a17e8a1a7d1cb41f1ecc5d7a042a2eb37a3ded082f828bd56939b1eae4f6884d" Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.848789 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckh9g"] Dec 15 12:20:55 crc kubenswrapper[4719]: I1215 12:20:55.860588 4719 scope.go:117] "RemoveContainer" containerID="d962fed45db1ba5c1d15f919342492d8fb6ba48f92b4b4bb7b5187a65260cbb6" Dec 15 12:20:57 crc kubenswrapper[4719]: I1215 12:20:57.489241 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" path="/var/lib/kubelet/pods/ea0b8649-9e2e-46d2-b9f0-5c51f155891e/volumes" Dec 15 12:20:58 crc kubenswrapper[4719]: I1215 12:20:58.352649 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:20:58 crc kubenswrapper[4719]: I1215 12:20:58.352739 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:20:58 crc kubenswrapper[4719]: I1215 12:20:58.408989 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:20:58 crc kubenswrapper[4719]: I1215 12:20:58.877420 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:20:59 crc kubenswrapper[4719]: I1215 12:20:59.611270 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:20:59 crc kubenswrapper[4719]: I1215 12:20:59.611581 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:20:59 crc kubenswrapper[4719]: I1215 12:20:59.655931 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:20:59 crc kubenswrapper[4719]: I1215 12:20:59.884959 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:21:06 crc kubenswrapper[4719]: I1215 12:21:06.964205 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkz2j"] Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.496156 4719 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.496808 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerName="extract-content" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.496824 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerName="extract-content" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.496840 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerName="extract-utilities" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.496925 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerName="extract-utilities" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.496945 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.496953 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.496965 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerName="extract-utilities" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.496974 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerName="extract-utilities" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.496989 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.496998 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.497008 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerName="extract-content" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497016 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerName="extract-content" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.497028 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerName="extract-content" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497036 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerName="extract-content" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.497049 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497057 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.497070 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerName="extract-content" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497078 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerName="extract-content" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.497090 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497098 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.497108 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerName="extract-utilities" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497116 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerName="extract-utilities" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.497130 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerName="extract-utilities" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497138 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerName="extract-utilities" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497280 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="b166625d-5aa1-46c1-9ec7-54aec2503fb7" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497295 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="6014e2e6-c315-4f83-9266-33239e0c43c4" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497313 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea0b8649-9e2e-46d2-b9f0-5c51f155891e" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497330 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f162837e-a473-4ee1-ac47-3caca3b12bbc" containerName="registry-server" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497694 4719 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.497895 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498045 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092" gracePeriod=15 Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498111 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2" gracePeriod=15 Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498172 4719 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498191 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e" gracePeriod=15 Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498127 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97" gracePeriod=15 Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498176 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02" gracePeriod=15 Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.498380 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498391 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.498403 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498408 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.498416 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498421 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.498428 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498433 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.498448 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498453 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.498463 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498468 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.498477 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498482 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498566 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498576 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498582 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498589 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498600 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.498606 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.504214 4719 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.538397 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.661495 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.661599 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.661630 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.661662 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.661690 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.661717 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.661750 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.661785 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763573 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763619 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763651 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763675 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763695 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763726 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763759 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763796 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763896 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763944 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763972 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.763999 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.764063 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.764103 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.764139 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.764170 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.833079 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:21:15 crc kubenswrapper[4719]: E1215 12:21:15.852820 4719 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188162e79f1855b6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-15 12:21:15.852084662 +0000 UTC m=+236.794377692,LastTimestamp:2025-12-15 12:21:15.852084662 +0000 UTC m=+236.794377692,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.940712 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"1a0f8820b8103cc0ae35ff17d2b2a1fb18dc04b91a6cb9bbd165f1c79a5d5771"} Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.942305 4719 generic.go:334] "Generic (PLEG): container finished" podID="6426eb80-427c-411c-bf47-af018703e000" containerID="0c58d556efb0decf61335e8d184efe3c5da120463bab307bab767c3acf82c75a" exitCode=0 Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.942341 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6426eb80-427c-411c-bf47-af018703e000","Type":"ContainerDied","Data":"0c58d556efb0decf61335e8d184efe3c5da120463bab307bab767c3acf82c75a"} Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.943039 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.943299 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.944242 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.945332 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.946233 4719 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97" exitCode=0 Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.946254 4719 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2" exitCode=0 Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.946261 4719 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02" exitCode=0 Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.946268 4719 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e" exitCode=2 Dec 15 12:21:15 crc kubenswrapper[4719]: I1215 12:21:15.946296 4719 scope.go:117] "RemoveContainer" containerID="f92212aeb4a09af22fc6de6b6cb733cb0b972e7b267aa7de26663e0741845bac" Dec 15 12:21:16 crc kubenswrapper[4719]: I1215 12:21:16.954134 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993"} Dec 15 12:21:16 crc kubenswrapper[4719]: I1215 12:21:16.955655 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:16 crc kubenswrapper[4719]: I1215 12:21:16.956083 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:16 crc kubenswrapper[4719]: I1215 12:21:16.959732 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.229633 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.230407 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.230747 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.385997 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-kubelet-dir\") pod \"6426eb80-427c-411c-bf47-af018703e000\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.386143 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-var-lock\") pod \"6426eb80-427c-411c-bf47-af018703e000\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.386170 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6426eb80-427c-411c-bf47-af018703e000-kube-api-access\") pod \"6426eb80-427c-411c-bf47-af018703e000\" (UID: \"6426eb80-427c-411c-bf47-af018703e000\") " Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.386688 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6426eb80-427c-411c-bf47-af018703e000" (UID: "6426eb80-427c-411c-bf47-af018703e000"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.386781 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-var-lock" (OuterVolumeSpecName: "var-lock") pod "6426eb80-427c-411c-bf47-af018703e000" (UID: "6426eb80-427c-411c-bf47-af018703e000"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.391964 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6426eb80-427c-411c-bf47-af018703e000-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6426eb80-427c-411c-bf47-af018703e000" (UID: "6426eb80-427c-411c-bf47-af018703e000"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.488131 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6426eb80-427c-411c-bf47-af018703e000-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.488190 4719 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-var-lock\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.488208 4719 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6426eb80-427c-411c-bf47-af018703e000-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.864599 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.865715 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.866318 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.866807 4719 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.867052 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.902818 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.902914 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.902918 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.902962 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.903025 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.903157 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.903188 4719 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.903208 4719 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.970292 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6426eb80-427c-411c-bf47-af018703e000","Type":"ContainerDied","Data":"eb61cca2433ad63dc066466de67e24311832bf1e0963c9ed0e85648cc7375fd4"} Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.970336 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb61cca2433ad63dc066466de67e24311832bf1e0963c9ed0e85648cc7375fd4" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.970404 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.977134 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.977958 4719 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.978692 4719 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092" exitCode=0 Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.978843 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.978988 4719 scope.go:117] "RemoveContainer" containerID="02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.981437 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:17 crc kubenswrapper[4719]: I1215 12:21:17.983151 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.000750 4719 scope.go:117] "RemoveContainer" containerID="39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.004340 4719 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.007606 4719 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.008113 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.008643 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.024484 4719 scope.go:117] "RemoveContainer" containerID="05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.042672 4719 scope.go:117] "RemoveContainer" containerID="063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.059108 4719 scope.go:117] "RemoveContainer" containerID="c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.080592 4719 scope.go:117] "RemoveContainer" containerID="4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.113925 4719 scope.go:117] "RemoveContainer" containerID="02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97" Dec 15 12:21:18 crc kubenswrapper[4719]: E1215 12:21:18.114448 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\": container with ID starting with 02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97 not found: ID does not exist" containerID="02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.114515 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97"} err="failed to get container status \"02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\": rpc error: code = NotFound desc = could not find container \"02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97\": container with ID starting with 02fae646f2642110802e176cd9da176d57b561a4fe3b271ab20b73032ad67e97 not found: ID does not exist" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.114568 4719 scope.go:117] "RemoveContainer" containerID="39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2" Dec 15 12:21:18 crc kubenswrapper[4719]: E1215 12:21:18.115088 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\": container with ID starting with 39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2 not found: ID does not exist" containerID="39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.115141 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2"} err="failed to get container status \"39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\": rpc error: code = NotFound desc = could not find container \"39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2\": container with ID starting with 39dc312d3e960908ae30250d1e448d551c3044982daecc5af84d86fbc67818c2 not found: ID does not exist" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.115179 4719 scope.go:117] "RemoveContainer" containerID="05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02" Dec 15 12:21:18 crc kubenswrapper[4719]: E1215 12:21:18.118156 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\": container with ID starting with 05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02 not found: ID does not exist" containerID="05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.118236 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02"} err="failed to get container status \"05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\": rpc error: code = NotFound desc = could not find container \"05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02\": container with ID starting with 05780f2aa6f2fb791f2b41ac5311e7fc3fababaec70761ce5e7d43319bf08d02 not found: ID does not exist" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.118291 4719 scope.go:117] "RemoveContainer" containerID="063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e" Dec 15 12:21:18 crc kubenswrapper[4719]: E1215 12:21:18.118689 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\": container with ID starting with 063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e not found: ID does not exist" containerID="063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.118762 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e"} err="failed to get container status \"063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\": rpc error: code = NotFound desc = could not find container \"063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e\": container with ID starting with 063de52a9630e52e389d9984c595d1eee7e0e8b14b42a08b234192e3ccd4de7e not found: ID does not exist" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.118802 4719 scope.go:117] "RemoveContainer" containerID="c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092" Dec 15 12:21:18 crc kubenswrapper[4719]: E1215 12:21:18.119211 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\": container with ID starting with c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092 not found: ID does not exist" containerID="c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.119272 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092"} err="failed to get container status \"c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\": rpc error: code = NotFound desc = could not find container \"c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092\": container with ID starting with c6a5a6c020fe7eb1b220ece62d8ae29da483e9dd545f567047ad3d43069ec092 not found: ID does not exist" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.119306 4719 scope.go:117] "RemoveContainer" containerID="4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650" Dec 15 12:21:18 crc kubenswrapper[4719]: E1215 12:21:18.119653 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\": container with ID starting with 4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650 not found: ID does not exist" containerID="4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650" Dec 15 12:21:18 crc kubenswrapper[4719]: I1215 12:21:18.119749 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650"} err="failed to get container status \"4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\": rpc error: code = NotFound desc = could not find container \"4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650\": container with ID starting with 4d7a1f36c2cf8e1643a4956f8487bb1fa55b28236e7cf43ad80eb8106d992650 not found: ID does not exist" Dec 15 12:21:19 crc kubenswrapper[4719]: I1215 12:21:19.480656 4719 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:19 crc kubenswrapper[4719]: I1215 12:21:19.481267 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:19 crc kubenswrapper[4719]: I1215 12:21:19.481776 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:19 crc kubenswrapper[4719]: I1215 12:21:19.483745 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 15 12:21:23 crc kubenswrapper[4719]: E1215 12:21:23.683453 4719 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:23 crc kubenswrapper[4719]: E1215 12:21:23.685524 4719 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:23 crc kubenswrapper[4719]: E1215 12:21:23.686154 4719 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:23 crc kubenswrapper[4719]: E1215 12:21:23.686762 4719 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:23 crc kubenswrapper[4719]: E1215 12:21:23.687382 4719 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:23 crc kubenswrapper[4719]: I1215 12:21:23.687451 4719 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 15 12:21:23 crc kubenswrapper[4719]: E1215 12:21:23.687946 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="200ms" Dec 15 12:21:23 crc kubenswrapper[4719]: E1215 12:21:23.889339 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="400ms" Dec 15 12:21:24 crc kubenswrapper[4719]: E1215 12:21:24.290743 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="800ms" Dec 15 12:21:24 crc kubenswrapper[4719]: E1215 12:21:24.796463 4719 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188162e79f1855b6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-15 12:21:15.852084662 +0000 UTC m=+236.794377692,LastTimestamp:2025-12-15 12:21:15.852084662 +0000 UTC m=+236.794377692,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 15 12:21:25 crc kubenswrapper[4719]: E1215 12:21:25.092430 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="1.6s" Dec 15 12:21:26 crc kubenswrapper[4719]: E1215 12:21:26.693329 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="3.2s" Dec 15 12:21:29 crc kubenswrapper[4719]: I1215 12:21:29.050933 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 15 12:21:29 crc kubenswrapper[4719]: I1215 12:21:29.052245 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054"} Dec 15 12:21:29 crc kubenswrapper[4719]: I1215 12:21:29.052144 4719 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054" exitCode=1 Dec 15 12:21:29 crc kubenswrapper[4719]: I1215 12:21:29.053130 4719 scope.go:117] "RemoveContainer" containerID="6ee45df8023ee1b6b1f4c638424ddb05629a937b8bc936092dbadc67f58b2054" Dec 15 12:21:29 crc kubenswrapper[4719]: I1215 12:21:29.053646 4719 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:29 crc kubenswrapper[4719]: I1215 12:21:29.054328 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:29 crc kubenswrapper[4719]: I1215 12:21:29.054849 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:29 crc kubenswrapper[4719]: I1215 12:21:29.482996 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:29 crc kubenswrapper[4719]: I1215 12:21:29.483487 4719 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:29 crc kubenswrapper[4719]: I1215 12:21:29.483934 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:29 crc kubenswrapper[4719]: E1215 12:21:29.895388 4719 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="6.4s" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.061578 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.061629 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"844da5edc16f3009d23e24a2cb99c768fcee480357363fa7e92557f9c82b2710"} Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.062369 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.062549 4719 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.062687 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.476143 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.478616 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.479295 4719 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.479845 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.499765 4719 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.499814 4719 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:21:30 crc kubenswrapper[4719]: E1215 12:21:30.500426 4719 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.501648 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:30 crc kubenswrapper[4719]: W1215 12:21:30.535148 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-c8f629d15a0d4a3df8e7ba8b6670027b44b2799269a3cd27479a08e1d0666fb1 WatchSource:0}: Error finding container c8f629d15a0d4a3df8e7ba8b6670027b44b2799269a3cd27479a08e1d0666fb1: Status 404 returned error can't find the container with id c8f629d15a0d4a3df8e7ba8b6670027b44b2799269a3cd27479a08e1d0666fb1 Dec 15 12:21:30 crc kubenswrapper[4719]: I1215 12:21:30.581148 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:21:31 crc kubenswrapper[4719]: I1215 12:21:31.278030 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c8f629d15a0d4a3df8e7ba8b6670027b44b2799269a3cd27479a08e1d0666fb1"} Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.016634 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" podUID="4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" containerName="oauth-openshift" containerID="cri-o://43d07ac5794200f80d34778ca3d8a1a55200521f904a19c956caa9a3d7542aa4" gracePeriod=15 Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.286585 4719 generic.go:334] "Generic (PLEG): container finished" podID="4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" containerID="43d07ac5794200f80d34778ca3d8a1a55200521f904a19c956caa9a3d7542aa4" exitCode=0 Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.286681 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" event={"ID":"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51","Type":"ContainerDied","Data":"43d07ac5794200f80d34778ca3d8a1a55200521f904a19c956caa9a3d7542aa4"} Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.288635 4719 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="de21b3c52d12f7242392c2eaeddf7c82605b2c7e18b412fb2e50b45e550f9ca0" exitCode=0 Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.288685 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"de21b3c52d12f7242392c2eaeddf7c82605b2c7e18b412fb2e50b45e550f9ca0"} Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.288964 4719 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.288979 4719 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.289647 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:32 crc kubenswrapper[4719]: E1215 12:21:32.289702 4719 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.289923 4719 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.290194 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.390026 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.390810 4719 status_manager.go:851] "Failed to get status for pod" podUID="4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-nkz2j\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.391540 4719 status_manager.go:851] "Failed to get status for pod" podUID="6426eb80-427c-411c-bf47-af018703e000" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.392295 4719 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.392639 4719 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.217:6443: connect: connection refused" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.588781 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-error\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589236 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-idp-0-file-data\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589286 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-dir\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589329 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-ocp-branding-template\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589379 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnlxk\" (UniqueName: \"kubernetes.io/projected/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-kube-api-access-jnlxk\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589414 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-trusted-ca-bundle\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589448 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-provider-selection\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589511 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-service-ca\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589554 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-policies\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589596 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-login\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589627 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-session\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589675 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-serving-cert\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589712 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-cliconfig\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.589781 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-router-certs\") pod \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\" (UID: \"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51\") " Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.590128 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.590785 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.590893 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.591464 4719 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.591501 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.591522 4719 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.596064 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.596962 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.597056 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-kube-api-access-jnlxk" (OuterVolumeSpecName: "kube-api-access-jnlxk") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "kube-api-access-jnlxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.597163 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.597467 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.597480 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.597524 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.602001 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.602247 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.603746 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.605044 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" (UID: "4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.691888 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.691922 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.691934 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.691944 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnlxk\" (UniqueName: \"kubernetes.io/projected/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-kube-api-access-jnlxk\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.691953 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.691962 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.691974 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.691983 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.691995 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.692005 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:32 crc kubenswrapper[4719]: I1215 12:21:32.692015 4719 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:21:33 crc kubenswrapper[4719]: I1215 12:21:33.301596 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ff402b39791f4b1b20134ebbfabd95dfa9d64454729f163804c9c2a782e49c09"} Dec 15 12:21:33 crc kubenswrapper[4719]: I1215 12:21:33.301940 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2ec60470cd108ea87b02f8e263e0b68e138645dfe25b2b394173cf7df80ff8bd"} Dec 15 12:21:33 crc kubenswrapper[4719]: I1215 12:21:33.301952 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"749b04a9916af7bd0832ccb672f85104ab606da8ae4d4896e1591946296a57e9"} Dec 15 12:21:33 crc kubenswrapper[4719]: I1215 12:21:33.301970 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5bfae332e5a967bbb9d40c4b48c08176ede8895976520f6f069034263a94b1b7"} Dec 15 12:21:33 crc kubenswrapper[4719]: I1215 12:21:33.303095 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" event={"ID":"4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51","Type":"ContainerDied","Data":"27b1509b041d1fba741ed2ce4e953aba4d79b33af8ae9fcce3fefcc390056477"} Dec 15 12:21:33 crc kubenswrapper[4719]: I1215 12:21:33.303124 4719 scope.go:117] "RemoveContainer" containerID="43d07ac5794200f80d34778ca3d8a1a55200521f904a19c956caa9a3d7542aa4" Dec 15 12:21:33 crc kubenswrapper[4719]: I1215 12:21:33.303317 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nkz2j" Dec 15 12:21:34 crc kubenswrapper[4719]: I1215 12:21:34.309762 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dbec2776a42dc9a46fe94ff1517fe0a85d5eca27673a0b3446111d8678534844"} Dec 15 12:21:34 crc kubenswrapper[4719]: I1215 12:21:34.310439 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:34 crc kubenswrapper[4719]: I1215 12:21:34.310655 4719 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:21:34 crc kubenswrapper[4719]: I1215 12:21:34.310776 4719 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:21:35 crc kubenswrapper[4719]: I1215 12:21:35.502796 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:35 crc kubenswrapper[4719]: I1215 12:21:35.502888 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:35 crc kubenswrapper[4719]: I1215 12:21:35.511107 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:36 crc kubenswrapper[4719]: I1215 12:21:36.356729 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:21:36 crc kubenswrapper[4719]: I1215 12:21:36.360171 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:21:39 crc kubenswrapper[4719]: I1215 12:21:39.319149 4719 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:39 crc kubenswrapper[4719]: I1215 12:21:39.494698 4719 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="6c2f606e-198e-4ee1-b893-cc3087273176" Dec 15 12:21:40 crc kubenswrapper[4719]: I1215 12:21:40.343891 4719 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:21:40 crc kubenswrapper[4719]: I1215 12:21:40.344351 4719 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:21:40 crc kubenswrapper[4719]: I1215 12:21:40.348375 4719 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="6c2f606e-198e-4ee1-b893-cc3087273176" Dec 15 12:21:40 crc kubenswrapper[4719]: I1215 12:21:40.351487 4719 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://5bfae332e5a967bbb9d40c4b48c08176ede8895976520f6f069034263a94b1b7" Dec 15 12:21:40 crc kubenswrapper[4719]: I1215 12:21:40.351676 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:21:40 crc kubenswrapper[4719]: I1215 12:21:40.584175 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 15 12:21:41 crc kubenswrapper[4719]: I1215 12:21:41.349636 4719 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:21:41 crc kubenswrapper[4719]: I1215 12:21:41.349671 4719 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:21:41 crc kubenswrapper[4719]: I1215 12:21:41.353548 4719 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="6c2f606e-198e-4ee1-b893-cc3087273176" Dec 15 12:21:48 crc kubenswrapper[4719]: I1215 12:21:48.859806 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 15 12:21:49 crc kubenswrapper[4719]: I1215 12:21:49.036060 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 15 12:21:49 crc kubenswrapper[4719]: I1215 12:21:49.262977 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 15 12:21:49 crc kubenswrapper[4719]: I1215 12:21:49.281462 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 15 12:21:49 crc kubenswrapper[4719]: I1215 12:21:49.338272 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 15 12:21:49 crc kubenswrapper[4719]: I1215 12:21:49.443705 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 15 12:21:49 crc kubenswrapper[4719]: I1215 12:21:49.468743 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 15 12:21:49 crc kubenswrapper[4719]: I1215 12:21:49.732316 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 15 12:21:50 crc kubenswrapper[4719]: I1215 12:21:50.137018 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 15 12:21:50 crc kubenswrapper[4719]: I1215 12:21:50.339506 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 15 12:21:50 crc kubenswrapper[4719]: I1215 12:21:50.409752 4719 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 15 12:21:50 crc kubenswrapper[4719]: I1215 12:21:50.581606 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 15 12:21:50 crc kubenswrapper[4719]: I1215 12:21:50.655967 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 15 12:21:50 crc kubenswrapper[4719]: I1215 12:21:50.715155 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 15 12:21:50 crc kubenswrapper[4719]: I1215 12:21:50.784223 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 15 12:21:50 crc kubenswrapper[4719]: I1215 12:21:50.980217 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.055731 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.124715 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.202765 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.207155 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.210963 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.246538 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.365241 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.380325 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.469380 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.525331 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.637717 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 15 12:21:51 crc kubenswrapper[4719]: I1215 12:21:51.697561 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.039064 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.058484 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.103094 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.126809 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.163482 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.208315 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.225686 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.280849 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.382238 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.425473 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.474506 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.520451 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.555138 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.646404 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.707055 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.751690 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 15 12:21:52 crc kubenswrapper[4719]: I1215 12:21:52.956101 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.076699 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.099141 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.142107 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.150091 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.179044 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.424208 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.463017 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.645272 4719 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.748362 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.766000 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.813606 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.898718 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.906066 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.930773 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 15 12:21:53 crc kubenswrapper[4719]: I1215 12:21:53.935285 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.012940 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.013936 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.065426 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.070493 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.070930 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.088846 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.091220 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.126601 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.289482 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.363151 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.372579 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.511226 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.516981 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.519565 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.614383 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.625725 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.795143 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.868205 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 15 12:21:54 crc kubenswrapper[4719]: I1215 12:21:54.951553 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.050052 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.078203 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.084335 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.091143 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.092246 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.150889 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.163187 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.242453 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.247653 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.306170 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.414773 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.442613 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.475235 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.893905 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 15 12:21:55 crc kubenswrapper[4719]: I1215 12:21:55.975152 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.036229 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.067211 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.078319 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.133690 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.195299 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.320702 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.320936 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.373877 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.543209 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.711236 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.729530 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.794774 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.832487 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.853974 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.855840 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 15 12:21:56 crc kubenswrapper[4719]: I1215 12:21:56.901959 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.048765 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.172690 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.250718 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.289201 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.367071 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.379501 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.394422 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.519240 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.569985 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.587963 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.614078 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.654958 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.763903 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.892615 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.914602 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 15 12:21:57 crc kubenswrapper[4719]: I1215 12:21:57.915515 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.035801 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.097155 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.108806 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.165627 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.186326 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.275209 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.376617 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.460153 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.647445 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.735910 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.747995 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.792294 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.868212 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.868409 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.874137 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.929153 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.939174 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 15 12:21:58 crc kubenswrapper[4719]: I1215 12:21:58.950432 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.223794 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.280712 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.486363 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.593614 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.595276 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.643337 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.652585 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.679413 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.727687 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.752395 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.767226 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.811518 4719 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.869221 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 15 12:21:59 crc kubenswrapper[4719]: I1215 12:21:59.887771 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.015011 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.172648 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.190956 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.239182 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.285521 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.347484 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.398665 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.437537 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.482353 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.512615 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.582578 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.601882 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.635394 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.637560 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.643040 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.790090 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.799633 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.834085 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.848842 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.862380 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.882981 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.890489 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.908545 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 15 12:22:00 crc kubenswrapper[4719]: I1215 12:22:00.979062 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.003238 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.097849 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.117645 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.144487 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.283458 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.460801 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.524009 4719 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.527330 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=46.527312089 podStartE2EDuration="46.527312089s" podCreationTimestamp="2025-12-15 12:21:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:21:38.931743986 +0000 UTC m=+259.874037016" watchObservedRunningTime="2025-12-15 12:22:01.527312089 +0000 UTC m=+282.469605119" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.528290 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkz2j","openshift-kube-apiserver/kube-apiserver-crc"] Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.528338 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh"] Dec 15 12:22:01 crc kubenswrapper[4719]: E1215 12:22:01.528537 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" containerName="oauth-openshift" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.528560 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" containerName="oauth-openshift" Dec 15 12:22:01 crc kubenswrapper[4719]: E1215 12:22:01.528589 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6426eb80-427c-411c-bf47-af018703e000" containerName="installer" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.528599 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6426eb80-427c-411c-bf47-af018703e000" containerName="installer" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.528690 4719 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.528717 4719 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a18fe895-3c85-44eb-b44d-54b25a8cc130" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.528720 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" containerName="oauth-openshift" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.528735 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="6426eb80-427c-411c-bf47-af018703e000" containerName="installer" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.529395 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.534171 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.534561 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.534571 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.536033 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.536518 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.537106 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.537116 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.537169 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.537110 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.537277 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.538826 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.538973 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.539167 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.552401 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.554748 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.557841 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.588325 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.58829998 podStartE2EDuration="22.58829998s" podCreationTimestamp="2025-12-15 12:21:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:22:01.565489464 +0000 UTC m=+282.507782494" watchObservedRunningTime="2025-12-15 12:22:01.58829998 +0000 UTC m=+282.530593030" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.598298 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.629841 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.659715 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.659757 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.659787 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9e14cefe-f576-4848-bc71-b1f597b0922d-audit-dir\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.659805 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.659821 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-template-login\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.659845 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.659883 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-service-ca\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.659927 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85r9z\" (UniqueName: \"kubernetes.io/projected/9e14cefe-f576-4848-bc71-b1f597b0922d-kube-api-access-85r9z\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.659943 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-audit-policies\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.659957 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.660617 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.660678 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-session\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.660704 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-router-certs\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.660738 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-template-error\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.761850 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-service-ca\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.761931 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85r9z\" (UniqueName: \"kubernetes.io/projected/9e14cefe-f576-4848-bc71-b1f597b0922d-kube-api-access-85r9z\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.761966 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-audit-policies\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.761988 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.762015 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.762057 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-session\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.762075 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-router-certs\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.762096 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-template-error\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.762118 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.762136 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.762159 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9e14cefe-f576-4848-bc71-b1f597b0922d-audit-dir\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.762173 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.762186 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-template-login\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.762206 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.763085 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-audit-policies\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.763787 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-service-ca\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.763957 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.765326 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9e14cefe-f576-4848-bc71-b1f597b0922d-audit-dir\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.767519 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.768487 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.769074 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.769449 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-router-certs\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.770026 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.770397 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-template-error\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.770632 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-template-login\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.771151 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.779346 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9e14cefe-f576-4848-bc71-b1f597b0922d-v4-0-config-system-session\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.783844 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85r9z\" (UniqueName: \"kubernetes.io/projected/9e14cefe-f576-4848-bc71-b1f597b0922d-kube-api-access-85r9z\") pod \"oauth-openshift-d8dcdf7ff-4zcvh\" (UID: \"9e14cefe-f576-4848-bc71-b1f597b0922d\") " pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.822625 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.852345 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.897252 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.903541 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.934829 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 15 12:22:01 crc kubenswrapper[4719]: I1215 12:22:01.972483 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.001600 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.061374 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.123447 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh"] Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.138715 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.217561 4719 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.348159 4719 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.463119 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" event={"ID":"9e14cefe-f576-4848-bc71-b1f597b0922d","Type":"ContainerStarted","Data":"93cb96524a91640c8506addb38d172fb37a9ded9fe6e71422e7e7cab307fbcd5"} Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.463169 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" event={"ID":"9e14cefe-f576-4848-bc71-b1f597b0922d","Type":"ContainerStarted","Data":"f4c44bf867aa382999d24a99d6edaf7bebf0ed854fc60a56c7329ca6c8f98b1a"} Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.463508 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.481774 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" podStartSLOduration=56.481755451 podStartE2EDuration="56.481755451s" podCreationTimestamp="2025-12-15 12:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:22:02.47961622 +0000 UTC m=+283.421909250" watchObservedRunningTime="2025-12-15 12:22:02.481755451 +0000 UTC m=+283.424048481" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.485993 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.616242 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.657201 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.673762 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-d8dcdf7ff-4zcvh" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.700105 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.903984 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 15 12:22:02 crc kubenswrapper[4719]: I1215 12:22:02.937368 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 15 12:22:03 crc kubenswrapper[4719]: I1215 12:22:03.234253 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 15 12:22:03 crc kubenswrapper[4719]: I1215 12:22:03.251136 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 15 12:22:03 crc kubenswrapper[4719]: I1215 12:22:03.356101 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 15 12:22:03 crc kubenswrapper[4719]: I1215 12:22:03.358486 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 15 12:22:03 crc kubenswrapper[4719]: I1215 12:22:03.383728 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 15 12:22:03 crc kubenswrapper[4719]: I1215 12:22:03.482768 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51" path="/var/lib/kubelet/pods/4a5ebe99-3de4-44dd-9bfa-bcc5473f9b51/volumes" Dec 15 12:22:03 crc kubenswrapper[4719]: I1215 12:22:03.697601 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 15 12:22:03 crc kubenswrapper[4719]: I1215 12:22:03.720615 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 15 12:22:03 crc kubenswrapper[4719]: I1215 12:22:03.739578 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 15 12:22:04 crc kubenswrapper[4719]: I1215 12:22:04.235386 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 15 12:22:04 crc kubenswrapper[4719]: I1215 12:22:04.856476 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 15 12:22:04 crc kubenswrapper[4719]: I1215 12:22:04.996079 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 15 12:22:05 crc kubenswrapper[4719]: I1215 12:22:05.125782 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 15 12:22:05 crc kubenswrapper[4719]: I1215 12:22:05.209017 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 15 12:22:05 crc kubenswrapper[4719]: I1215 12:22:05.236065 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 15 12:22:05 crc kubenswrapper[4719]: I1215 12:22:05.283950 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 15 12:22:12 crc kubenswrapper[4719]: I1215 12:22:12.899198 4719 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 15 12:22:12 crc kubenswrapper[4719]: I1215 12:22:12.899938 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993" gracePeriod=5 Dec 15 12:22:13 crc kubenswrapper[4719]: I1215 12:22:13.520697 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 15 12:22:16 crc kubenswrapper[4719]: I1215 12:22:16.557223 4719 generic.go:334] "Generic (PLEG): container finished" podID="556c5cc5-d57e-413d-aa79-421d196cb799" containerID="cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121" exitCode=0 Dec 15 12:22:16 crc kubenswrapper[4719]: I1215 12:22:16.557417 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" event={"ID":"556c5cc5-d57e-413d-aa79-421d196cb799","Type":"ContainerDied","Data":"cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121"} Dec 15 12:22:16 crc kubenswrapper[4719]: I1215 12:22:16.557989 4719 scope.go:117] "RemoveContainer" containerID="cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121" Dec 15 12:22:17 crc kubenswrapper[4719]: I1215 12:22:17.563671 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" event={"ID":"556c5cc5-d57e-413d-aa79-421d196cb799","Type":"ContainerStarted","Data":"32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e"} Dec 15 12:22:17 crc kubenswrapper[4719]: I1215 12:22:17.564933 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:22:17 crc kubenswrapper[4719]: I1215 12:22:17.567435 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.461374 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.461903 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.567259 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.567361 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.568176 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.568464 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.568598 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.568728 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.568370 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.568488 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.568643 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.569312 4719 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.569419 4719 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.569507 4719 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.569598 4719 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.571507 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.571553 4719 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993" exitCode=137 Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.571643 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.571657 4719 scope.go:117] "RemoveContainer" containerID="7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.579318 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.608164 4719 scope.go:117] "RemoveContainer" containerID="7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993" Dec 15 12:22:18 crc kubenswrapper[4719]: E1215 12:22:18.613161 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993\": container with ID starting with 7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993 not found: ID does not exist" containerID="7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.613205 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993"} err="failed to get container status \"7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993\": rpc error: code = NotFound desc = could not find container \"7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993\": container with ID starting with 7911c5993a481003067d1e3d720a4c7bdeac510cb32436f4281e21fe591d2993 not found: ID does not exist" Dec 15 12:22:18 crc kubenswrapper[4719]: I1215 12:22:18.670690 4719 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:19 crc kubenswrapper[4719]: I1215 12:22:19.350434 4719 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 15 12:22:19 crc kubenswrapper[4719]: I1215 12:22:19.498306 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 15 12:22:19 crc kubenswrapper[4719]: I1215 12:22:19.498983 4719 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 15 12:22:19 crc kubenswrapper[4719]: I1215 12:22:19.510830 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 15 12:22:19 crc kubenswrapper[4719]: I1215 12:22:19.510946 4719 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="ce7c42c5-11d7-4994-9f5e-ee320c9a8808" Dec 15 12:22:19 crc kubenswrapper[4719]: I1215 12:22:19.514740 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 15 12:22:19 crc kubenswrapper[4719]: I1215 12:22:19.514765 4719 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="ce7c42c5-11d7-4994-9f5e-ee320c9a8808" Dec 15 12:22:21 crc kubenswrapper[4719]: I1215 12:22:21.825218 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 15 12:22:22 crc kubenswrapper[4719]: I1215 12:22:22.834094 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 15 12:22:26 crc kubenswrapper[4719]: I1215 12:22:26.063452 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 15 12:22:26 crc kubenswrapper[4719]: I1215 12:22:26.993930 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 15 12:22:27 crc kubenswrapper[4719]: I1215 12:22:27.139948 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 15 12:22:28 crc kubenswrapper[4719]: I1215 12:22:28.172101 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 15 12:22:29 crc kubenswrapper[4719]: I1215 12:22:29.618200 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 15 12:22:30 crc kubenswrapper[4719]: I1215 12:22:30.840418 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 15 12:22:30 crc kubenswrapper[4719]: I1215 12:22:30.857790 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 15 12:22:30 crc kubenswrapper[4719]: I1215 12:22:30.970925 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 15 12:22:31 crc kubenswrapper[4719]: I1215 12:22:31.491330 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 15 12:22:32 crc kubenswrapper[4719]: I1215 12:22:32.216258 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 15 12:22:33 crc kubenswrapper[4719]: I1215 12:22:33.615404 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nz9wh"] Dec 15 12:22:33 crc kubenswrapper[4719]: I1215 12:22:33.615967 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" podUID="e6665e73-ac68-4fcc-a0e0-85815184a0ea" containerName="controller-manager" containerID="cri-o://a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67" gracePeriod=30 Dec 15 12:22:33 crc kubenswrapper[4719]: I1215 12:22:33.743392 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr"] Dec 15 12:22:33 crc kubenswrapper[4719]: I1215 12:22:33.743799 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" podUID="5fe9cf63-1287-418f-b2c5-4e4e7458cf44" containerName="route-controller-manager" containerID="cri-o://f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788" gracePeriod=30 Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.038283 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.090491 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.160332 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-client-ca\") pod \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.160393 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6665e73-ac68-4fcc-a0e0-85815184a0ea-serving-cert\") pod \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.160426 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-config\") pod \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.160465 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrv8b\" (UniqueName: \"kubernetes.io/projected/e6665e73-ac68-4fcc-a0e0-85815184a0ea-kube-api-access-nrv8b\") pod \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.160522 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-proxy-ca-bundles\") pod \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\" (UID: \"e6665e73-ac68-4fcc-a0e0-85815184a0ea\") " Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.161193 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e6665e73-ac68-4fcc-a0e0-85815184a0ea" (UID: "e6665e73-ac68-4fcc-a0e0-85815184a0ea"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.161586 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-client-ca" (OuterVolumeSpecName: "client-ca") pod "e6665e73-ac68-4fcc-a0e0-85815184a0ea" (UID: "e6665e73-ac68-4fcc-a0e0-85815184a0ea"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.161807 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-config" (OuterVolumeSpecName: "config") pod "e6665e73-ac68-4fcc-a0e0-85815184a0ea" (UID: "e6665e73-ac68-4fcc-a0e0-85815184a0ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.165221 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6665e73-ac68-4fcc-a0e0-85815184a0ea-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e6665e73-ac68-4fcc-a0e0-85815184a0ea" (UID: "e6665e73-ac68-4fcc-a0e0-85815184a0ea"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.165446 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6665e73-ac68-4fcc-a0e0-85815184a0ea-kube-api-access-nrv8b" (OuterVolumeSpecName: "kube-api-access-nrv8b") pod "e6665e73-ac68-4fcc-a0e0-85815184a0ea" (UID: "e6665e73-ac68-4fcc-a0e0-85815184a0ea"). InnerVolumeSpecName "kube-api-access-nrv8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.261170 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-config\") pod \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.261218 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hp2g\" (UniqueName: \"kubernetes.io/projected/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-kube-api-access-2hp2g\") pod \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.261251 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-client-ca\") pod \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.261339 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-serving-cert\") pod \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\" (UID: \"5fe9cf63-1287-418f-b2c5-4e4e7458cf44\") " Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.262007 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-client-ca" (OuterVolumeSpecName: "client-ca") pod "5fe9cf63-1287-418f-b2c5-4e4e7458cf44" (UID: "5fe9cf63-1287-418f-b2c5-4e4e7458cf44"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.262154 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-config" (OuterVolumeSpecName: "config") pod "5fe9cf63-1287-418f-b2c5-4e4e7458cf44" (UID: "5fe9cf63-1287-418f-b2c5-4e4e7458cf44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.262498 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6665e73-ac68-4fcc-a0e0-85815184a0ea-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.262518 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.262528 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrv8b\" (UniqueName: \"kubernetes.io/projected/e6665e73-ac68-4fcc-a0e0-85815184a0ea-kube-api-access-nrv8b\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.262539 4719 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.262547 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.262575 4719 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-client-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.262585 4719 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6665e73-ac68-4fcc-a0e0-85815184a0ea-client-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.265092 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5fe9cf63-1287-418f-b2c5-4e4e7458cf44" (UID: "5fe9cf63-1287-418f-b2c5-4e4e7458cf44"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.265143 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-kube-api-access-2hp2g" (OuterVolumeSpecName: "kube-api-access-2hp2g") pod "5fe9cf63-1287-418f-b2c5-4e4e7458cf44" (UID: "5fe9cf63-1287-418f-b2c5-4e4e7458cf44"). InnerVolumeSpecName "kube-api-access-2hp2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.363315 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hp2g\" (UniqueName: \"kubernetes.io/projected/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-kube-api-access-2hp2g\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.363366 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe9cf63-1287-418f-b2c5-4e4e7458cf44-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.678354 4719 generic.go:334] "Generic (PLEG): container finished" podID="5fe9cf63-1287-418f-b2c5-4e4e7458cf44" containerID="f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788" exitCode=0 Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.678421 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.678465 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" event={"ID":"5fe9cf63-1287-418f-b2c5-4e4e7458cf44","Type":"ContainerDied","Data":"f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788"} Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.678497 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr" event={"ID":"5fe9cf63-1287-418f-b2c5-4e4e7458cf44","Type":"ContainerDied","Data":"aae65cef2246ccafec9cf563c241796d847fae667522399f3442c503e8aeef4c"} Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.678518 4719 scope.go:117] "RemoveContainer" containerID="f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.680333 4719 generic.go:334] "Generic (PLEG): container finished" podID="e6665e73-ac68-4fcc-a0e0-85815184a0ea" containerID="a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67" exitCode=0 Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.680374 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" event={"ID":"e6665e73-ac68-4fcc-a0e0-85815184a0ea","Type":"ContainerDied","Data":"a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67"} Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.680386 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.680402 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nz9wh" event={"ID":"e6665e73-ac68-4fcc-a0e0-85815184a0ea","Type":"ContainerDied","Data":"65a7c1c640aced6bbb321c169e97f419b80ceaa9fe61ae6c575ca7f7e61bed92"} Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.699711 4719 scope.go:117] "RemoveContainer" containerID="f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788" Dec 15 12:22:34 crc kubenswrapper[4719]: E1215 12:22:34.700369 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788\": container with ID starting with f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788 not found: ID does not exist" containerID="f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.700421 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788"} err="failed to get container status \"f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788\": rpc error: code = NotFound desc = could not find container \"f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788\": container with ID starting with f770b702720918a41ae39f92076c0c3eb552e00ddc5a1f0850ede39ee9068788 not found: ID does not exist" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.700456 4719 scope.go:117] "RemoveContainer" containerID="a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.712275 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr"] Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.719359 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sm6zr"] Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.728956 4719 scope.go:117] "RemoveContainer" containerID="a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67" Dec 15 12:22:34 crc kubenswrapper[4719]: E1215 12:22:34.729375 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67\": container with ID starting with a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67 not found: ID does not exist" containerID="a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.729413 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67"} err="failed to get container status \"a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67\": rpc error: code = NotFound desc = could not find container \"a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67\": container with ID starting with a73014d17eb9edef7707e2bcb5807fc24bb850b28eb3cabfdcac628c4f72be67 not found: ID does not exist" Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.736770 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nz9wh"] Dec 15 12:22:34 crc kubenswrapper[4719]: I1215 12:22:34.742294 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nz9wh"] Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.191073 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.483287 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe9cf63-1287-418f-b2c5-4e4e7458cf44" path="/var/lib/kubelet/pods/5fe9cf63-1287-418f-b2c5-4e4e7458cf44/volumes" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.484229 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6665e73-ac68-4fcc-a0e0-85815184a0ea" path="/var/lib/kubelet/pods/e6665e73-ac68-4fcc-a0e0-85815184a0ea/volumes" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.509187 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk"] Dec 15 12:22:35 crc kubenswrapper[4719]: E1215 12:22:35.510497 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe9cf63-1287-418f-b2c5-4e4e7458cf44" containerName="route-controller-manager" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.510551 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe9cf63-1287-418f-b2c5-4e4e7458cf44" containerName="route-controller-manager" Dec 15 12:22:35 crc kubenswrapper[4719]: E1215 12:22:35.510580 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.510589 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 15 12:22:35 crc kubenswrapper[4719]: E1215 12:22:35.510601 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6665e73-ac68-4fcc-a0e0-85815184a0ea" containerName="controller-manager" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.510609 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6665e73-ac68-4fcc-a0e0-85815184a0ea" containerName="controller-manager" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.510875 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6665e73-ac68-4fcc-a0e0-85815184a0ea" containerName="controller-manager" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.510891 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.510905 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe9cf63-1287-418f-b2c5-4e4e7458cf44" containerName="route-controller-manager" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.511353 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.514263 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz"] Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.515096 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.515111 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.516794 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.517338 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.522090 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.522119 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.525740 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.525752 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.525955 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.526288 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.526325 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.526760 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.526909 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.531714 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.533973 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz"] Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.546127 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk"] Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.577038 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e255f8eb-ae56-4da0-a721-2775ff5c597c-serving-cert\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.577620 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e255f8eb-ae56-4da0-a721-2775ff5c597c-config\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.577698 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e255f8eb-ae56-4da0-a721-2775ff5c597c-client-ca\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.678262 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e255f8eb-ae56-4da0-a721-2775ff5c597c-serving-cert\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.678330 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-proxy-ca-bundles\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.678361 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e255f8eb-ae56-4da0-a721-2775ff5c597c-config\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.678379 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e255f8eb-ae56-4da0-a721-2775ff5c597c-client-ca\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.678407 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a54763b3-8252-4fef-9cca-eff62eca5280-serving-cert\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.678424 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-config\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.678454 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-client-ca\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.678484 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkpps\" (UniqueName: \"kubernetes.io/projected/a54763b3-8252-4fef-9cca-eff62eca5280-kube-api-access-qkpps\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.678511 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stqtd\" (UniqueName: \"kubernetes.io/projected/e255f8eb-ae56-4da0-a721-2775ff5c597c-kube-api-access-stqtd\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.680051 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e255f8eb-ae56-4da0-a721-2775ff5c597c-config\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.680214 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e255f8eb-ae56-4da0-a721-2775ff5c597c-client-ca\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.683818 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e255f8eb-ae56-4da0-a721-2775ff5c597c-serving-cert\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.779449 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a54763b3-8252-4fef-9cca-eff62eca5280-serving-cert\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.779501 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-config\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.779537 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-client-ca\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.779554 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkpps\" (UniqueName: \"kubernetes.io/projected/a54763b3-8252-4fef-9cca-eff62eca5280-kube-api-access-qkpps\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.779583 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stqtd\" (UniqueName: \"kubernetes.io/projected/e255f8eb-ae56-4da0-a721-2775ff5c597c-kube-api-access-stqtd\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.779611 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-proxy-ca-bundles\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.781128 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-proxy-ca-bundles\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.782194 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-config\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.782910 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-client-ca\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.784080 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a54763b3-8252-4fef-9cca-eff62eca5280-serving-cert\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.797416 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkpps\" (UniqueName: \"kubernetes.io/projected/a54763b3-8252-4fef-9cca-eff62eca5280-kube-api-access-qkpps\") pod \"controller-manager-55cf5b8bc-ljrkz\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.797414 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stqtd\" (UniqueName: \"kubernetes.io/projected/e255f8eb-ae56-4da0-a721-2775ff5c597c-kube-api-access-stqtd\") pod \"route-controller-manager-5954d7f6fc-xh7kk\" (UID: \"e255f8eb-ae56-4da0-a721-2775ff5c597c\") " pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.833241 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:35 crc kubenswrapper[4719]: I1215 12:22:35.846491 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.023737 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk"] Dec 15 12:22:36 crc kubenswrapper[4719]: W1215 12:22:36.038352 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode255f8eb_ae56_4da0_a721_2775ff5c597c.slice/crio-fad2c9d5da0dd55d762574faf188932a42d6432fae9f6a345461cabe75aa6aac WatchSource:0}: Error finding container fad2c9d5da0dd55d762574faf188932a42d6432fae9f6a345461cabe75aa6aac: Status 404 returned error can't find the container with id fad2c9d5da0dd55d762574faf188932a42d6432fae9f6a345461cabe75aa6aac Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.269703 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz"] Dec 15 12:22:36 crc kubenswrapper[4719]: W1215 12:22:36.277597 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda54763b3_8252_4fef_9cca_eff62eca5280.slice/crio-59a2476e878838e32c9f9b2d79f693621428a1720511dac8b301f77162f9d2df WatchSource:0}: Error finding container 59a2476e878838e32c9f9b2d79f693621428a1720511dac8b301f77162f9d2df: Status 404 returned error can't find the container with id 59a2476e878838e32c9f9b2d79f693621428a1720511dac8b301f77162f9d2df Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.694502 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" event={"ID":"a54763b3-8252-4fef-9cca-eff62eca5280","Type":"ContainerStarted","Data":"51eee20c71a05d83a0edcaf4af2e8ebc4d5ad1944af4747116d2e47da3faef3d"} Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.694551 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" event={"ID":"a54763b3-8252-4fef-9cca-eff62eca5280","Type":"ContainerStarted","Data":"59a2476e878838e32c9f9b2d79f693621428a1720511dac8b301f77162f9d2df"} Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.694741 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.696260 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" event={"ID":"e255f8eb-ae56-4da0-a721-2775ff5c597c","Type":"ContainerStarted","Data":"753020d11cb3c074cfb81589ec61e62984f7d9b9daa9dd05bf97ea580108cc5e"} Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.696292 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" event={"ID":"e255f8eb-ae56-4da0-a721-2775ff5c597c","Type":"ContainerStarted","Data":"fad2c9d5da0dd55d762574faf188932a42d6432fae9f6a345461cabe75aa6aac"} Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.696489 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.701979 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.715318 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" podStartSLOduration=3.7152950750000002 podStartE2EDuration="3.715295075s" podCreationTimestamp="2025-12-15 12:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:22:36.711736788 +0000 UTC m=+317.654029818" watchObservedRunningTime="2025-12-15 12:22:36.715295075 +0000 UTC m=+317.657588105" Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.720369 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" Dec 15 12:22:36 crc kubenswrapper[4719]: I1215 12:22:36.733017 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5954d7f6fc-xh7kk" podStartSLOduration=3.733000622 podStartE2EDuration="3.733000622s" podCreationTimestamp="2025-12-15 12:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:22:36.730017993 +0000 UTC m=+317.672311043" watchObservedRunningTime="2025-12-15 12:22:36.733000622 +0000 UTC m=+317.675293652" Dec 15 12:22:43 crc kubenswrapper[4719]: I1215 12:22:43.980776 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 15 12:22:51 crc kubenswrapper[4719]: I1215 12:22:51.357291 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:22:51 crc kubenswrapper[4719]: I1215 12:22:51.358282 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:22:53 crc kubenswrapper[4719]: I1215 12:22:53.606599 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz"] Dec 15 12:22:53 crc kubenswrapper[4719]: I1215 12:22:53.606844 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" podUID="a54763b3-8252-4fef-9cca-eff62eca5280" containerName="controller-manager" containerID="cri-o://51eee20c71a05d83a0edcaf4af2e8ebc4d5ad1944af4747116d2e47da3faef3d" gracePeriod=30 Dec 15 12:22:53 crc kubenswrapper[4719]: I1215 12:22:53.828141 4719 generic.go:334] "Generic (PLEG): container finished" podID="a54763b3-8252-4fef-9cca-eff62eca5280" containerID="51eee20c71a05d83a0edcaf4af2e8ebc4d5ad1944af4747116d2e47da3faef3d" exitCode=0 Dec 15 12:22:53 crc kubenswrapper[4719]: I1215 12:22:53.828242 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" event={"ID":"a54763b3-8252-4fef-9cca-eff62eca5280","Type":"ContainerDied","Data":"51eee20c71a05d83a0edcaf4af2e8ebc4d5ad1944af4747116d2e47da3faef3d"} Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.143972 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.216935 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-config\") pod \"a54763b3-8252-4fef-9cca-eff62eca5280\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.217051 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-proxy-ca-bundles\") pod \"a54763b3-8252-4fef-9cca-eff62eca5280\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.217882 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-config" (OuterVolumeSpecName: "config") pod "a54763b3-8252-4fef-9cca-eff62eca5280" (UID: "a54763b3-8252-4fef-9cca-eff62eca5280"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.218109 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a54763b3-8252-4fef-9cca-eff62eca5280" (UID: "a54763b3-8252-4fef-9cca-eff62eca5280"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.218185 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a54763b3-8252-4fef-9cca-eff62eca5280-serving-cert\") pod \"a54763b3-8252-4fef-9cca-eff62eca5280\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.219010 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-client-ca\") pod \"a54763b3-8252-4fef-9cca-eff62eca5280\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.219098 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkpps\" (UniqueName: \"kubernetes.io/projected/a54763b3-8252-4fef-9cca-eff62eca5280-kube-api-access-qkpps\") pod \"a54763b3-8252-4fef-9cca-eff62eca5280\" (UID: \"a54763b3-8252-4fef-9cca-eff62eca5280\") " Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.219359 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-client-ca" (OuterVolumeSpecName: "client-ca") pod "a54763b3-8252-4fef-9cca-eff62eca5280" (UID: "a54763b3-8252-4fef-9cca-eff62eca5280"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.219981 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.220002 4719 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.220013 4719 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a54763b3-8252-4fef-9cca-eff62eca5280-client-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.223129 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a54763b3-8252-4fef-9cca-eff62eca5280-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a54763b3-8252-4fef-9cca-eff62eca5280" (UID: "a54763b3-8252-4fef-9cca-eff62eca5280"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.223373 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a54763b3-8252-4fef-9cca-eff62eca5280-kube-api-access-qkpps" (OuterVolumeSpecName: "kube-api-access-qkpps") pod "a54763b3-8252-4fef-9cca-eff62eca5280" (UID: "a54763b3-8252-4fef-9cca-eff62eca5280"). InnerVolumeSpecName "kube-api-access-qkpps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.320907 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkpps\" (UniqueName: \"kubernetes.io/projected/a54763b3-8252-4fef-9cca-eff62eca5280-kube-api-access-qkpps\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.320998 4719 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a54763b3-8252-4fef-9cca-eff62eca5280-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.835052 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" event={"ID":"a54763b3-8252-4fef-9cca-eff62eca5280","Type":"ContainerDied","Data":"59a2476e878838e32c9f9b2d79f693621428a1720511dac8b301f77162f9d2df"} Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.835086 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.835397 4719 scope.go:117] "RemoveContainer" containerID="51eee20c71a05d83a0edcaf4af2e8ebc4d5ad1944af4747116d2e47da3faef3d" Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.883947 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz"] Dec 15 12:22:54 crc kubenswrapper[4719]: I1215 12:22:54.891051 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-55cf5b8bc-ljrkz"] Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.488020 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a54763b3-8252-4fef-9cca-eff62eca5280" path="/var/lib/kubelet/pods/a54763b3-8252-4fef-9cca-eff62eca5280/volumes" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.519153 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-79889b6c87-zkhmg"] Dec 15 12:22:55 crc kubenswrapper[4719]: E1215 12:22:55.519395 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54763b3-8252-4fef-9cca-eff62eca5280" containerName="controller-manager" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.519411 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54763b3-8252-4fef-9cca-eff62eca5280" containerName="controller-manager" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.519550 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="a54763b3-8252-4fef-9cca-eff62eca5280" containerName="controller-manager" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.520005 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.521909 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.522674 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.522888 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.523054 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.523851 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.525154 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.533328 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.536416 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20911566-51ec-4a99-813e-7098446b9828-proxy-ca-bundles\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.536472 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20911566-51ec-4a99-813e-7098446b9828-client-ca\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.536524 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20911566-51ec-4a99-813e-7098446b9828-serving-cert\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.536572 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20911566-51ec-4a99-813e-7098446b9828-config\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.536596 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f624n\" (UniqueName: \"kubernetes.io/projected/20911566-51ec-4a99-813e-7098446b9828-kube-api-access-f624n\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.539172 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79889b6c87-zkhmg"] Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.637465 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20911566-51ec-4a99-813e-7098446b9828-client-ca\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.637520 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20911566-51ec-4a99-813e-7098446b9828-serving-cert\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.637576 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20911566-51ec-4a99-813e-7098446b9828-config\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.637594 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f624n\" (UniqueName: \"kubernetes.io/projected/20911566-51ec-4a99-813e-7098446b9828-kube-api-access-f624n\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.637709 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20911566-51ec-4a99-813e-7098446b9828-proxy-ca-bundles\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.638391 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20911566-51ec-4a99-813e-7098446b9828-client-ca\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.638976 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20911566-51ec-4a99-813e-7098446b9828-proxy-ca-bundles\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.639613 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20911566-51ec-4a99-813e-7098446b9828-config\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.641616 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20911566-51ec-4a99-813e-7098446b9828-serving-cert\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.658389 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f624n\" (UniqueName: \"kubernetes.io/projected/20911566-51ec-4a99-813e-7098446b9828-kube-api-access-f624n\") pod \"controller-manager-79889b6c87-zkhmg\" (UID: \"20911566-51ec-4a99-813e-7098446b9828\") " pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:55 crc kubenswrapper[4719]: I1215 12:22:55.847334 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:56 crc kubenswrapper[4719]: I1215 12:22:56.229603 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79889b6c87-zkhmg"] Dec 15 12:22:56 crc kubenswrapper[4719]: W1215 12:22:56.243064 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20911566_51ec_4a99_813e_7098446b9828.slice/crio-5cdfc00f8601d742ebc2b79a823d6a6455e4e6778cf5028f03ae64c51d0bba9f WatchSource:0}: Error finding container 5cdfc00f8601d742ebc2b79a823d6a6455e4e6778cf5028f03ae64c51d0bba9f: Status 404 returned error can't find the container with id 5cdfc00f8601d742ebc2b79a823d6a6455e4e6778cf5028f03ae64c51d0bba9f Dec 15 12:22:56 crc kubenswrapper[4719]: I1215 12:22:56.848714 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" event={"ID":"20911566-51ec-4a99-813e-7098446b9828","Type":"ContainerStarted","Data":"300a9db6f6623cbda8588d04a892b3377eeb9cf213e087ab84f1afc8ac772df0"} Dec 15 12:22:56 crc kubenswrapper[4719]: I1215 12:22:56.848765 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" event={"ID":"20911566-51ec-4a99-813e-7098446b9828","Type":"ContainerStarted","Data":"5cdfc00f8601d742ebc2b79a823d6a6455e4e6778cf5028f03ae64c51d0bba9f"} Dec 15 12:22:56 crc kubenswrapper[4719]: I1215 12:22:56.848995 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:56 crc kubenswrapper[4719]: I1215 12:22:56.853441 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" Dec 15 12:22:56 crc kubenswrapper[4719]: I1215 12:22:56.886186 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-79889b6c87-zkhmg" podStartSLOduration=3.886167038 podStartE2EDuration="3.886167038s" podCreationTimestamp="2025-12-15 12:22:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:22:56.868121463 +0000 UTC m=+337.810414493" watchObservedRunningTime="2025-12-15 12:22:56.886167038 +0000 UTC m=+337.828460068" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.173950 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6279r"] Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.175242 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.221238 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6279r"] Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.301205 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slvqz\" (UniqueName: \"kubernetes.io/projected/4f498f97-c302-4378-a570-38b545d1ee15-kube-api-access-slvqz\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.301256 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.301278 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f498f97-c302-4378-a570-38b545d1ee15-trusted-ca\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.301298 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4f498f97-c302-4378-a570-38b545d1ee15-registry-certificates\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.301326 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4f498f97-c302-4378-a570-38b545d1ee15-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.301354 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4f498f97-c302-4378-a570-38b545d1ee15-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.301383 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4f498f97-c302-4378-a570-38b545d1ee15-bound-sa-token\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.301412 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4f498f97-c302-4378-a570-38b545d1ee15-registry-tls\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.331263 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.402616 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4f498f97-c302-4378-a570-38b545d1ee15-bound-sa-token\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.402675 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4f498f97-c302-4378-a570-38b545d1ee15-registry-tls\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.402714 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slvqz\" (UniqueName: \"kubernetes.io/projected/4f498f97-c302-4378-a570-38b545d1ee15-kube-api-access-slvqz\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.402745 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f498f97-c302-4378-a570-38b545d1ee15-trusted-ca\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.402770 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4f498f97-c302-4378-a570-38b545d1ee15-registry-certificates\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.402802 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4f498f97-c302-4378-a570-38b545d1ee15-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.402833 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4f498f97-c302-4378-a570-38b545d1ee15-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.404297 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f498f97-c302-4378-a570-38b545d1ee15-trusted-ca\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.404443 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4f498f97-c302-4378-a570-38b545d1ee15-registry-certificates\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.404504 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4f498f97-c302-4378-a570-38b545d1ee15-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.412269 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4f498f97-c302-4378-a570-38b545d1ee15-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.415719 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4f498f97-c302-4378-a570-38b545d1ee15-registry-tls\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.422380 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4f498f97-c302-4378-a570-38b545d1ee15-bound-sa-token\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.429079 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slvqz\" (UniqueName: \"kubernetes.io/projected/4f498f97-c302-4378-a570-38b545d1ee15-kube-api-access-slvqz\") pod \"image-registry-66df7c8f76-6279r\" (UID: \"4f498f97-c302-4378-a570-38b545d1ee15\") " pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.490999 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.890003 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6279r"] Dec 15 12:23:09 crc kubenswrapper[4719]: I1215 12:23:09.916288 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6279r" event={"ID":"4f498f97-c302-4378-a570-38b545d1ee15","Type":"ContainerStarted","Data":"3117807dc8b099e2da513736a17d22bebb4f32bbf2774df3ec637d9456810740"} Dec 15 12:23:10 crc kubenswrapper[4719]: I1215 12:23:10.923249 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6279r" event={"ID":"4f498f97-c302-4378-a570-38b545d1ee15","Type":"ContainerStarted","Data":"894f875d372d2fe34b70ae31c61873d999111ecb5319d5f17229cee7b0c3fef0"} Dec 15 12:23:10 crc kubenswrapper[4719]: I1215 12:23:10.924398 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:21 crc kubenswrapper[4719]: I1215 12:23:21.357365 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:23:21 crc kubenswrapper[4719]: I1215 12:23:21.358051 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:23:29 crc kubenswrapper[4719]: I1215 12:23:29.495898 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-6279r" Dec 15 12:23:29 crc kubenswrapper[4719]: I1215 12:23:29.525127 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-6279r" podStartSLOduration=20.525099232 podStartE2EDuration="20.525099232s" podCreationTimestamp="2025-12-15 12:23:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:23:10.95553374 +0000 UTC m=+351.897826780" watchObservedRunningTime="2025-12-15 12:23:29.525099232 +0000 UTC m=+370.467392292" Dec 15 12:23:29 crc kubenswrapper[4719]: I1215 12:23:29.554044 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gfmg4"] Dec 15 12:23:51 crc kubenswrapper[4719]: I1215 12:23:51.356497 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:23:51 crc kubenswrapper[4719]: I1215 12:23:51.357144 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:23:51 crc kubenswrapper[4719]: I1215 12:23:51.357196 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:23:51 crc kubenswrapper[4719]: I1215 12:23:51.357635 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"38e050f4b3eb5ec1de6c0f4f194704b5cba328b77e0bccea20d90d6077185480"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 12:23:51 crc kubenswrapper[4719]: I1215 12:23:51.357705 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://38e050f4b3eb5ec1de6c0f4f194704b5cba328b77e0bccea20d90d6077185480" gracePeriod=600 Dec 15 12:23:52 crc kubenswrapper[4719]: I1215 12:23:52.169593 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="38e050f4b3eb5ec1de6c0f4f194704b5cba328b77e0bccea20d90d6077185480" exitCode=0 Dec 15 12:23:52 crc kubenswrapper[4719]: I1215 12:23:52.169762 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"38e050f4b3eb5ec1de6c0f4f194704b5cba328b77e0bccea20d90d6077185480"} Dec 15 12:23:52 crc kubenswrapper[4719]: I1215 12:23:52.169846 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"d92d6e0171c7041e83e55a8b1507b34e41f6fa135abdaf671e5b50dd4ed3cf24"} Dec 15 12:23:52 crc kubenswrapper[4719]: I1215 12:23:52.169894 4719 scope.go:117] "RemoveContainer" containerID="080b95a6b39cd38ee1265166043a8c7041bf40a67ad47b2155f77b64b4df92a9" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.139564 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vqndj"] Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.140796 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vqndj" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerName="registry-server" containerID="cri-o://5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594" gracePeriod=30 Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.151988 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2xhs"] Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.152267 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k2xhs" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerName="registry-server" containerID="cri-o://abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960" gracePeriod=30 Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.161974 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t6qkh"] Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.162210 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" containerName="marketplace-operator" containerID="cri-o://32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e" gracePeriod=30 Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.179760 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kh6zf"] Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.180035 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kh6zf" podUID="20547bc3-5b4d-40fd-898e-49b148705327" containerName="registry-server" containerID="cri-o://2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877" gracePeriod=30 Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.192041 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ds7r2"] Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.192793 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.196369 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gwjsp"] Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.196584 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gwjsp" podUID="638821b0-deb7-40ee-bad1-b552994b4598" containerName="registry-server" containerID="cri-o://8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f" gracePeriod=30 Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.223240 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ds7r2"] Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.377357 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ds7r2\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.377415 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ds7r2\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.377450 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brnw5\" (UniqueName: \"kubernetes.io/projected/5de2277b-38a8-4be5-b4f5-b7516197192c-kube-api-access-brnw5\") pod \"marketplace-operator-79b997595-ds7r2\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.478500 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ds7r2\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.478759 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ds7r2\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.478786 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brnw5\" (UniqueName: \"kubernetes.io/projected/5de2277b-38a8-4be5-b4f5-b7516197192c-kube-api-access-brnw5\") pod \"marketplace-operator-79b997595-ds7r2\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.479787 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ds7r2\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.497765 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ds7r2\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.508829 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brnw5\" (UniqueName: \"kubernetes.io/projected/5de2277b-38a8-4be5-b4f5-b7516197192c-kube-api-access-brnw5\") pod \"marketplace-operator-79b997595-ds7r2\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.514727 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.634552 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.688718 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-catalog-content\") pod \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.693630 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-utilities" (OuterVolumeSpecName: "utilities") pod "722c74db-78bf-49c2-8ff7-03cba8ea3afc" (UID: "722c74db-78bf-49c2-8ff7-03cba8ea3afc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.693750 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.693961 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-utilities\") pod \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.694966 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzjhp\" (UniqueName: \"kubernetes.io/projected/722c74db-78bf-49c2-8ff7-03cba8ea3afc-kube-api-access-lzjhp\") pod \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\" (UID: \"722c74db-78bf-49c2-8ff7-03cba8ea3afc\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.695316 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.702016 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/722c74db-78bf-49c2-8ff7-03cba8ea3afc-kube-api-access-lzjhp" (OuterVolumeSpecName: "kube-api-access-lzjhp") pod "722c74db-78bf-49c2-8ff7-03cba8ea3afc" (UID: "722c74db-78bf-49c2-8ff7-03cba8ea3afc"). InnerVolumeSpecName "kube-api-access-lzjhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.756337 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.767078 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "722c74db-78bf-49c2-8ff7-03cba8ea3afc" (UID: "722c74db-78bf-49c2-8ff7-03cba8ea3afc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.784210 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.795895 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-trusted-ca\") pod \"556c5cc5-d57e-413d-aa79-421d196cb799\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.795940 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvhtz\" (UniqueName: \"kubernetes.io/projected/638821b0-deb7-40ee-bad1-b552994b4598-kube-api-access-hvhtz\") pod \"638821b0-deb7-40ee-bad1-b552994b4598\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.795964 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-catalog-content\") pod \"638821b0-deb7-40ee-bad1-b552994b4598\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.796017 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb7kh\" (UniqueName: \"kubernetes.io/projected/556c5cc5-d57e-413d-aa79-421d196cb799-kube-api-access-fb7kh\") pod \"556c5cc5-d57e-413d-aa79-421d196cb799\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.796053 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-utilities\") pod \"638821b0-deb7-40ee-bad1-b552994b4598\" (UID: \"638821b0-deb7-40ee-bad1-b552994b4598\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.796079 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-operator-metrics\") pod \"556c5cc5-d57e-413d-aa79-421d196cb799\" (UID: \"556c5cc5-d57e-413d-aa79-421d196cb799\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.796266 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzjhp\" (UniqueName: \"kubernetes.io/projected/722c74db-78bf-49c2-8ff7-03cba8ea3afc-kube-api-access-lzjhp\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.796277 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/722c74db-78bf-49c2-8ff7-03cba8ea3afc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.798126 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "556c5cc5-d57e-413d-aa79-421d196cb799" (UID: "556c5cc5-d57e-413d-aa79-421d196cb799"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.799116 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-utilities" (OuterVolumeSpecName: "utilities") pod "638821b0-deb7-40ee-bad1-b552994b4598" (UID: "638821b0-deb7-40ee-bad1-b552994b4598"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.801754 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/556c5cc5-d57e-413d-aa79-421d196cb799-kube-api-access-fb7kh" (OuterVolumeSpecName: "kube-api-access-fb7kh") pod "556c5cc5-d57e-413d-aa79-421d196cb799" (UID: "556c5cc5-d57e-413d-aa79-421d196cb799"). InnerVolumeSpecName "kube-api-access-fb7kh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.802293 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "556c5cc5-d57e-413d-aa79-421d196cb799" (UID: "556c5cc5-d57e-413d-aa79-421d196cb799"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.802314 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/638821b0-deb7-40ee-bad1-b552994b4598-kube-api-access-hvhtz" (OuterVolumeSpecName: "kube-api-access-hvhtz") pod "638821b0-deb7-40ee-bad1-b552994b4598" (UID: "638821b0-deb7-40ee-bad1-b552994b4598"). InnerVolumeSpecName "kube-api-access-hvhtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.811741 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.897052 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbjvv\" (UniqueName: \"kubernetes.io/projected/cdd254ce-5006-40e1-a7f8-0850d049242f-kube-api-access-xbjvv\") pod \"cdd254ce-5006-40e1-a7f8-0850d049242f\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.897133 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-catalog-content\") pod \"20547bc3-5b4d-40fd-898e-49b148705327\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.897174 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-utilities\") pod \"20547bc3-5b4d-40fd-898e-49b148705327\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.897211 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-catalog-content\") pod \"cdd254ce-5006-40e1-a7f8-0850d049242f\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.897227 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-utilities\") pod \"cdd254ce-5006-40e1-a7f8-0850d049242f\" (UID: \"cdd254ce-5006-40e1-a7f8-0850d049242f\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.897245 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbdq8\" (UniqueName: \"kubernetes.io/projected/20547bc3-5b4d-40fd-898e-49b148705327-kube-api-access-cbdq8\") pod \"20547bc3-5b4d-40fd-898e-49b148705327\" (UID: \"20547bc3-5b4d-40fd-898e-49b148705327\") " Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.897821 4719 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.898632 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-utilities" (OuterVolumeSpecName: "utilities") pod "cdd254ce-5006-40e1-a7f8-0850d049242f" (UID: "cdd254ce-5006-40e1-a7f8-0850d049242f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.898685 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-utilities" (OuterVolumeSpecName: "utilities") pod "20547bc3-5b4d-40fd-898e-49b148705327" (UID: "20547bc3-5b4d-40fd-898e-49b148705327"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.900691 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20547bc3-5b4d-40fd-898e-49b148705327-kube-api-access-cbdq8" (OuterVolumeSpecName: "kube-api-access-cbdq8") pod "20547bc3-5b4d-40fd-898e-49b148705327" (UID: "20547bc3-5b4d-40fd-898e-49b148705327"). InnerVolumeSpecName "kube-api-access-cbdq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.903412 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdd254ce-5006-40e1-a7f8-0850d049242f-kube-api-access-xbjvv" (OuterVolumeSpecName: "kube-api-access-xbjvv") pod "cdd254ce-5006-40e1-a7f8-0850d049242f" (UID: "cdd254ce-5006-40e1-a7f8-0850d049242f"). InnerVolumeSpecName "kube-api-access-xbjvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.897836 4719 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/556c5cc5-d57e-413d-aa79-421d196cb799-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.903956 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvhtz\" (UniqueName: \"kubernetes.io/projected/638821b0-deb7-40ee-bad1-b552994b4598-kube-api-access-hvhtz\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.903974 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb7kh\" (UniqueName: \"kubernetes.io/projected/556c5cc5-d57e-413d-aa79-421d196cb799-kube-api-access-fb7kh\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.903990 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.927037 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20547bc3-5b4d-40fd-898e-49b148705327" (UID: "20547bc3-5b4d-40fd-898e-49b148705327"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.930527 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ds7r2"] Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.958455 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "638821b0-deb7-40ee-bad1-b552994b4598" (UID: "638821b0-deb7-40ee-bad1-b552994b4598"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:23:53 crc kubenswrapper[4719]: I1215 12:23:53.987594 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cdd254ce-5006-40e1-a7f8-0850d049242f" (UID: "cdd254ce-5006-40e1-a7f8-0850d049242f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.005075 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.005116 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.005127 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd254ce-5006-40e1-a7f8-0850d049242f-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.005136 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbdq8\" (UniqueName: \"kubernetes.io/projected/20547bc3-5b4d-40fd-898e-49b148705327-kube-api-access-cbdq8\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.005145 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638821b0-deb7-40ee-bad1-b552994b4598-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.005154 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbjvv\" (UniqueName: \"kubernetes.io/projected/cdd254ce-5006-40e1-a7f8-0850d049242f-kube-api-access-xbjvv\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.005162 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20547bc3-5b4d-40fd-898e-49b148705327-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.182098 4719 generic.go:334] "Generic (PLEG): container finished" podID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerID="5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594" exitCode=0 Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.182175 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqndj" event={"ID":"722c74db-78bf-49c2-8ff7-03cba8ea3afc","Type":"ContainerDied","Data":"5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.182212 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vqndj" event={"ID":"722c74db-78bf-49c2-8ff7-03cba8ea3afc","Type":"ContainerDied","Data":"c1761e87867d14dea8fbbc05f4a38dd13d99cd5e6bcf4562061db1f861bc0ba5"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.182232 4719 scope.go:117] "RemoveContainer" containerID="5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.182214 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vqndj" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.184474 4719 generic.go:334] "Generic (PLEG): container finished" podID="638821b0-deb7-40ee-bad1-b552994b4598" containerID="8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f" exitCode=0 Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.184569 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gwjsp" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.185051 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gwjsp" event={"ID":"638821b0-deb7-40ee-bad1-b552994b4598","Type":"ContainerDied","Data":"8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.185081 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gwjsp" event={"ID":"638821b0-deb7-40ee-bad1-b552994b4598","Type":"ContainerDied","Data":"eeac450a182307fd5f5f3a9f17bece47aea0dd411f9bcd4c98cefe259ba80d2a"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.189197 4719 generic.go:334] "Generic (PLEG): container finished" podID="20547bc3-5b4d-40fd-898e-49b148705327" containerID="2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877" exitCode=0 Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.189255 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kh6zf" event={"ID":"20547bc3-5b4d-40fd-898e-49b148705327","Type":"ContainerDied","Data":"2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.189278 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kh6zf" event={"ID":"20547bc3-5b4d-40fd-898e-49b148705327","Type":"ContainerDied","Data":"c75f6e32799777d9187f19d09dded38af1d7f485f274b54fcc58653deab58e00"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.189358 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kh6zf" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.197715 4719 scope.go:117] "RemoveContainer" containerID="346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.197918 4719 generic.go:334] "Generic (PLEG): container finished" podID="556c5cc5-d57e-413d-aa79-421d196cb799" containerID="32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e" exitCode=0 Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.197973 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" event={"ID":"556c5cc5-d57e-413d-aa79-421d196cb799","Type":"ContainerDied","Data":"32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.198009 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" event={"ID":"556c5cc5-d57e-413d-aa79-421d196cb799","Type":"ContainerDied","Data":"2082cbdee92bbf92f31b7999981f3ea9b4a566f877d23e4329895c981ce1d428"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.198081 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t6qkh" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.200908 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" event={"ID":"5de2277b-38a8-4be5-b4f5-b7516197192c","Type":"ContainerStarted","Data":"3a6d10cfe2ade7410f3ac7ea893a64ff7cb491f5ec4e015f333f7eedd228f538"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.200944 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" event={"ID":"5de2277b-38a8-4be5-b4f5-b7516197192c","Type":"ContainerStarted","Data":"c38b86ca2e0f66f1b37faeb4e328170f3d894a950a26f1142130b29dcf3d3e09"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.200959 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.204977 4719 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ds7r2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" start-of-body= Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.205006 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" podUID="5de2277b-38a8-4be5-b4f5-b7516197192c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.210649 4719 generic.go:334] "Generic (PLEG): container finished" podID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerID="abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960" exitCode=0 Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.210688 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2xhs" event={"ID":"cdd254ce-5006-40e1-a7f8-0850d049242f","Type":"ContainerDied","Data":"abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.210710 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k2xhs" event={"ID":"cdd254ce-5006-40e1-a7f8-0850d049242f","Type":"ContainerDied","Data":"dd3f41b2b96471a1030d748e6f98950385bde3ecb089646e180092de0a0504e6"} Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.210918 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k2xhs" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.236363 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" podStartSLOduration=1.236339877 podStartE2EDuration="1.236339877s" podCreationTimestamp="2025-12-15 12:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:23:54.226642917 +0000 UTC m=+395.168935947" watchObservedRunningTime="2025-12-15 12:23:54.236339877 +0000 UTC m=+395.178632917" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.247092 4719 scope.go:117] "RemoveContainer" containerID="9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.250847 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gwjsp"] Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.256953 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gwjsp"] Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.262312 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vqndj"] Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.268407 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vqndj"] Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.269839 4719 scope.go:117] "RemoveContainer" containerID="5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.272195 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594\": container with ID starting with 5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594 not found: ID does not exist" containerID="5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.272233 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594"} err="failed to get container status \"5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594\": rpc error: code = NotFound desc = could not find container \"5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594\": container with ID starting with 5ef9749f54385f69b793d2b5759f4440e7caae3cc8b319ec856382a9f7724594 not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.272257 4719 scope.go:117] "RemoveContainer" containerID="346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.276015 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94\": container with ID starting with 346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94 not found: ID does not exist" containerID="346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.276080 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94"} err="failed to get container status \"346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94\": rpc error: code = NotFound desc = could not find container \"346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94\": container with ID starting with 346944e9a33bc8d71cae71d7506c9a47cadd5912901ee21b92f9d40fc4d47f94 not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.276129 4719 scope.go:117] "RemoveContainer" containerID="9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.276543 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c\": container with ID starting with 9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c not found: ID does not exist" containerID="9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.276577 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c"} err="failed to get container status \"9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c\": rpc error: code = NotFound desc = could not find container \"9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c\": container with ID starting with 9d325cc99ba37b7a6854a2dc9f281ed40c23b13656db0fde3a4f788d58fd392c not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.276608 4719 scope.go:117] "RemoveContainer" containerID="8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.285499 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kh6zf"] Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.306809 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kh6zf"] Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.317762 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t6qkh"] Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.323185 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t6qkh"] Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.334247 4719 scope.go:117] "RemoveContainer" containerID="4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.335187 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k2xhs"] Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.338405 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k2xhs"] Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.349845 4719 scope.go:117] "RemoveContainer" containerID="c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.364876 4719 scope.go:117] "RemoveContainer" containerID="8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.365480 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f\": container with ID starting with 8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f not found: ID does not exist" containerID="8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.365523 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f"} err="failed to get container status \"8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f\": rpc error: code = NotFound desc = could not find container \"8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f\": container with ID starting with 8bf8aac4fb994db7fce66dce49aa08d44ff9ccc260ab9bf1ab47436627bd610f not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.365555 4719 scope.go:117] "RemoveContainer" containerID="4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.366284 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53\": container with ID starting with 4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53 not found: ID does not exist" containerID="4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.366310 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53"} err="failed to get container status \"4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53\": rpc error: code = NotFound desc = could not find container \"4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53\": container with ID starting with 4daf1114163187361e8c7dee05e04fb1f557ce7508ccb5828e6b56f8c72d1e53 not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.366331 4719 scope.go:117] "RemoveContainer" containerID="c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.366618 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a\": container with ID starting with c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a not found: ID does not exist" containerID="c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.366659 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a"} err="failed to get container status \"c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a\": rpc error: code = NotFound desc = could not find container \"c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a\": container with ID starting with c48266c1e22fe42dc0bb1b689f95bb5af1798dba46d4f43b95ce432a71f9e77a not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.366674 4719 scope.go:117] "RemoveContainer" containerID="2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.384753 4719 scope.go:117] "RemoveContainer" containerID="cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.398134 4719 scope.go:117] "RemoveContainer" containerID="196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.416819 4719 scope.go:117] "RemoveContainer" containerID="2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.417209 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877\": container with ID starting with 2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877 not found: ID does not exist" containerID="2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.417236 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877"} err="failed to get container status \"2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877\": rpc error: code = NotFound desc = could not find container \"2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877\": container with ID starting with 2d0504c22e9ee20dbdbb9b3588703b625df552b2f1adf0ca364d2742e26ec877 not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.417256 4719 scope.go:117] "RemoveContainer" containerID="cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.417632 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546\": container with ID starting with cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546 not found: ID does not exist" containerID="cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.417778 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546"} err="failed to get container status \"cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546\": rpc error: code = NotFound desc = could not find container \"cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546\": container with ID starting with cbe1643200a2b2217cbfd8d89884deb9e7c9b52689910b3cb9825c628d304546 not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.417914 4719 scope.go:117] "RemoveContainer" containerID="196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.418376 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0\": container with ID starting with 196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0 not found: ID does not exist" containerID="196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.418398 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0"} err="failed to get container status \"196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0\": rpc error: code = NotFound desc = could not find container \"196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0\": container with ID starting with 196bc3ad659d96c612096ed10b097e64f66ef2e6a4c497831ab51e20480382f0 not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.418415 4719 scope.go:117] "RemoveContainer" containerID="32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.433951 4719 scope.go:117] "RemoveContainer" containerID="cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.446572 4719 scope.go:117] "RemoveContainer" containerID="32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.446957 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e\": container with ID starting with 32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e not found: ID does not exist" containerID="32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.446981 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e"} err="failed to get container status \"32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e\": rpc error: code = NotFound desc = could not find container \"32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e\": container with ID starting with 32e98f2629d332edbf7499fbe1518f7209626b1bd5f31072a93a5b84f87b403e not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.446998 4719 scope.go:117] "RemoveContainer" containerID="cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.447183 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121\": container with ID starting with cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121 not found: ID does not exist" containerID="cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.447202 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121"} err="failed to get container status \"cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121\": rpc error: code = NotFound desc = could not find container \"cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121\": container with ID starting with cf03e8818f5999279ccbaaede4d7c63966a65fe4fa2f1dc9cc7caa105e97d121 not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.447215 4719 scope.go:117] "RemoveContainer" containerID="abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.462058 4719 scope.go:117] "RemoveContainer" containerID="00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.474755 4719 scope.go:117] "RemoveContainer" containerID="56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.486633 4719 scope.go:117] "RemoveContainer" containerID="abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.486878 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960\": container with ID starting with abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960 not found: ID does not exist" containerID="abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.486905 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960"} err="failed to get container status \"abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960\": rpc error: code = NotFound desc = could not find container \"abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960\": container with ID starting with abd44c451f5acf2a5a2be83b47e63f23357a8782448d69a7f0b5a30c08ade960 not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.486925 4719 scope.go:117] "RemoveContainer" containerID="00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.487317 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d\": container with ID starting with 00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d not found: ID does not exist" containerID="00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.487336 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d"} err="failed to get container status \"00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d\": rpc error: code = NotFound desc = could not find container \"00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d\": container with ID starting with 00f58f744a8d2c78233f4d52132560e9490bf25c37ed81eeb56b8c5c6a1af12d not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.487351 4719 scope.go:117] "RemoveContainer" containerID="56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84" Dec 15 12:23:54 crc kubenswrapper[4719]: E1215 12:23:54.487517 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84\": container with ID starting with 56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84 not found: ID does not exist" containerID="56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.487543 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84"} err="failed to get container status \"56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84\": rpc error: code = NotFound desc = could not find container \"56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84\": container with ID starting with 56aa5ee89eed458065c89a26985c338fb26a70b42b280a81038a9a4fa0665a84 not found: ID does not exist" Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.596111 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" podUID="27255302-b4b6-40dc-a5fd-f6e376b7e05d" containerName="registry" containerID="cri-o://c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124" gracePeriod=30 Dec 15 12:23:54 crc kubenswrapper[4719]: I1215 12:23:54.947037 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.017177 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-bound-sa-token\") pod \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.017461 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/27255302-b4b6-40dc-a5fd-f6e376b7e05d-installation-pull-secrets\") pod \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.017595 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.017625 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dqww\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-kube-api-access-6dqww\") pod \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.017644 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-trusted-ca\") pod \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.017661 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/27255302-b4b6-40dc-a5fd-f6e376b7e05d-ca-trust-extracted\") pod \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.017690 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-certificates\") pod \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.017709 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-tls\") pod \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\" (UID: \"27255302-b4b6-40dc-a5fd-f6e376b7e05d\") " Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.019374 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "27255302-b4b6-40dc-a5fd-f6e376b7e05d" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.019429 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "27255302-b4b6-40dc-a5fd-f6e376b7e05d" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.023900 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27255302-b4b6-40dc-a5fd-f6e376b7e05d-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "27255302-b4b6-40dc-a5fd-f6e376b7e05d" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.023942 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "27255302-b4b6-40dc-a5fd-f6e376b7e05d" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.024118 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-kube-api-access-6dqww" (OuterVolumeSpecName: "kube-api-access-6dqww") pod "27255302-b4b6-40dc-a5fd-f6e376b7e05d" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d"). InnerVolumeSpecName "kube-api-access-6dqww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.024646 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "27255302-b4b6-40dc-a5fd-f6e376b7e05d" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.028620 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "27255302-b4b6-40dc-a5fd-f6e376b7e05d" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.037063 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27255302-b4b6-40dc-a5fd-f6e376b7e05d-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "27255302-b4b6-40dc-a5fd-f6e376b7e05d" (UID: "27255302-b4b6-40dc-a5fd-f6e376b7e05d"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.118923 4719 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.119268 4719 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.119339 4719 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.119399 4719 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/27255302-b4b6-40dc-a5fd-f6e376b7e05d-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.119456 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dqww\" (UniqueName: \"kubernetes.io/projected/27255302-b4b6-40dc-a5fd-f6e376b7e05d-kube-api-access-6dqww\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.119517 4719 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/27255302-b4b6-40dc-a5fd-f6e376b7e05d-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.119573 4719 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/27255302-b4b6-40dc-a5fd-f6e376b7e05d-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.223788 4719 generic.go:334] "Generic (PLEG): container finished" podID="27255302-b4b6-40dc-a5fd-f6e376b7e05d" containerID="c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124" exitCode=0 Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.226244 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.229879 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" event={"ID":"27255302-b4b6-40dc-a5fd-f6e376b7e05d","Type":"ContainerDied","Data":"c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124"} Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.231754 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gfmg4" event={"ID":"27255302-b4b6-40dc-a5fd-f6e376b7e05d","Type":"ContainerDied","Data":"42a20bd63897c11e3e862c545d5b7d458dd2701956a4a1c6be8ed4bddeb150d3"} Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.231882 4719 scope.go:117] "RemoveContainer" containerID="c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.238627 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.250158 4719 scope.go:117] "RemoveContainer" containerID="c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.256625 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124\": container with ID starting with c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124 not found: ID does not exist" containerID="c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.256668 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124"} err="failed to get container status \"c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124\": rpc error: code = NotFound desc = could not find container \"c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124\": container with ID starting with c39e2e6816c83686a12fdb81a49c97634c8f96f22cfa6ac061625bba6f44d124 not found: ID does not exist" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.292601 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gfmg4"] Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.299171 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gfmg4"] Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359405 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fxvcb"] Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359662 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerName="extract-content" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359681 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerName="extract-content" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359695 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" containerName="marketplace-operator" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359703 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" containerName="marketplace-operator" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359716 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerName="extract-content" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359724 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerName="extract-content" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359737 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359746 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359756 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638821b0-deb7-40ee-bad1-b552994b4598" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359764 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="638821b0-deb7-40ee-bad1-b552994b4598" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359777 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638821b0-deb7-40ee-bad1-b552994b4598" containerName="extract-utilities" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359784 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="638821b0-deb7-40ee-bad1-b552994b4598" containerName="extract-utilities" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359795 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" containerName="marketplace-operator" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359807 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" containerName="marketplace-operator" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359817 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20547bc3-5b4d-40fd-898e-49b148705327" containerName="extract-content" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359827 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="20547bc3-5b4d-40fd-898e-49b148705327" containerName="extract-content" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359837 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359845 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359874 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerName="extract-utilities" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359882 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerName="extract-utilities" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359894 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20547bc3-5b4d-40fd-898e-49b148705327" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359901 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="20547bc3-5b4d-40fd-898e-49b148705327" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359910 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27255302-b4b6-40dc-a5fd-f6e376b7e05d" containerName="registry" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.359982 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="27255302-b4b6-40dc-a5fd-f6e376b7e05d" containerName="registry" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.359998 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638821b0-deb7-40ee-bad1-b552994b4598" containerName="extract-content" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.360007 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="638821b0-deb7-40ee-bad1-b552994b4598" containerName="extract-content" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.360019 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20547bc3-5b4d-40fd-898e-49b148705327" containerName="extract-utilities" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.360027 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="20547bc3-5b4d-40fd-898e-49b148705327" containerName="extract-utilities" Dec 15 12:23:55 crc kubenswrapper[4719]: E1215 12:23:55.360035 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerName="extract-utilities" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.360044 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerName="extract-utilities" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.360164 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="27255302-b4b6-40dc-a5fd-f6e376b7e05d" containerName="registry" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.360180 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" containerName="marketplace-operator" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.360191 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.360199 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.360208 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="638821b0-deb7-40ee-bad1-b552994b4598" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.360224 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="20547bc3-5b4d-40fd-898e-49b148705327" containerName="registry-server" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.362513 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" containerName="marketplace-operator" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.363780 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.366104 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxvcb"] Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.366417 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.422119 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-utilities\") pod \"redhat-marketplace-fxvcb\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.422161 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqj6v\" (UniqueName: \"kubernetes.io/projected/44695712-4b17-480e-8c22-a392046822ca-kube-api-access-jqj6v\") pod \"redhat-marketplace-fxvcb\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.422206 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-catalog-content\") pod \"redhat-marketplace-fxvcb\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.485235 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20547bc3-5b4d-40fd-898e-49b148705327" path="/var/lib/kubelet/pods/20547bc3-5b4d-40fd-898e-49b148705327/volumes" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.486728 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27255302-b4b6-40dc-a5fd-f6e376b7e05d" path="/var/lib/kubelet/pods/27255302-b4b6-40dc-a5fd-f6e376b7e05d/volumes" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.487648 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="556c5cc5-d57e-413d-aa79-421d196cb799" path="/var/lib/kubelet/pods/556c5cc5-d57e-413d-aa79-421d196cb799/volumes" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.489087 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="638821b0-deb7-40ee-bad1-b552994b4598" path="/var/lib/kubelet/pods/638821b0-deb7-40ee-bad1-b552994b4598/volumes" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.489948 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="722c74db-78bf-49c2-8ff7-03cba8ea3afc" path="/var/lib/kubelet/pods/722c74db-78bf-49c2-8ff7-03cba8ea3afc/volumes" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.491197 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdd254ce-5006-40e1-a7f8-0850d049242f" path="/var/lib/kubelet/pods/cdd254ce-5006-40e1-a7f8-0850d049242f/volumes" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.523549 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-catalog-content\") pod \"redhat-marketplace-fxvcb\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.523652 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-utilities\") pod \"redhat-marketplace-fxvcb\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.523671 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqj6v\" (UniqueName: \"kubernetes.io/projected/44695712-4b17-480e-8c22-a392046822ca-kube-api-access-jqj6v\") pod \"redhat-marketplace-fxvcb\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.524769 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-catalog-content\") pod \"redhat-marketplace-fxvcb\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.525288 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-utilities\") pod \"redhat-marketplace-fxvcb\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.555061 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqj6v\" (UniqueName: \"kubernetes.io/projected/44695712-4b17-480e-8c22-a392046822ca-kube-api-access-jqj6v\") pod \"redhat-marketplace-fxvcb\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.560872 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xl7hj"] Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.563749 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.567785 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.586666 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xl7hj"] Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.624948 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-utilities\") pod \"certified-operators-xl7hj\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.625063 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-catalog-content\") pod \"certified-operators-xl7hj\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.625161 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-584kg\" (UniqueName: \"kubernetes.io/projected/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-kube-api-access-584kg\") pod \"certified-operators-xl7hj\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.685068 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.726348 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-utilities\") pod \"certified-operators-xl7hj\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.726717 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-catalog-content\") pod \"certified-operators-xl7hj\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.727061 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-utilities\") pod \"certified-operators-xl7hj\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.727110 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-catalog-content\") pod \"certified-operators-xl7hj\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.727169 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-584kg\" (UniqueName: \"kubernetes.io/projected/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-kube-api-access-584kg\") pod \"certified-operators-xl7hj\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.745496 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-584kg\" (UniqueName: \"kubernetes.io/projected/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-kube-api-access-584kg\") pod \"certified-operators-xl7hj\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:55 crc kubenswrapper[4719]: I1215 12:23:55.896292 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:23:56 crc kubenswrapper[4719]: I1215 12:23:56.114693 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxvcb"] Dec 15 12:23:56 crc kubenswrapper[4719]: I1215 12:23:56.230954 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxvcb" event={"ID":"44695712-4b17-480e-8c22-a392046822ca","Type":"ContainerStarted","Data":"5701cbadfdb10fcc1aefb5fbff79e44a36bbdff4f3bb4916e9dab0c9c8bf96ca"} Dec 15 12:23:56 crc kubenswrapper[4719]: I1215 12:23:56.231359 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxvcb" event={"ID":"44695712-4b17-480e-8c22-a392046822ca","Type":"ContainerStarted","Data":"04d88408b05bf6fc2eece9f29bd6d2fbb544d5371732aa1a7f85276682149e17"} Dec 15 12:23:56 crc kubenswrapper[4719]: I1215 12:23:56.286804 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xl7hj"] Dec 15 12:23:56 crc kubenswrapper[4719]: W1215 12:23:56.296543 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06c7e7eb_5780_4b52_bd86_f5b74a20f8bc.slice/crio-b12997847ca460285bc1b20b78dcc282176b2b151c93435e5230fa3978e605bb WatchSource:0}: Error finding container b12997847ca460285bc1b20b78dcc282176b2b151c93435e5230fa3978e605bb: Status 404 returned error can't find the container with id b12997847ca460285bc1b20b78dcc282176b2b151c93435e5230fa3978e605bb Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.238432 4719 generic.go:334] "Generic (PLEG): container finished" podID="44695712-4b17-480e-8c22-a392046822ca" containerID="5701cbadfdb10fcc1aefb5fbff79e44a36bbdff4f3bb4916e9dab0c9c8bf96ca" exitCode=0 Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.238586 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxvcb" event={"ID":"44695712-4b17-480e-8c22-a392046822ca","Type":"ContainerDied","Data":"5701cbadfdb10fcc1aefb5fbff79e44a36bbdff4f3bb4916e9dab0c9c8bf96ca"} Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.242176 4719 generic.go:334] "Generic (PLEG): container finished" podID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerID="f9e8561e08ec1866df76a77520de09f0d2cb296c59e50127f48fe2ad0217d4fb" exitCode=0 Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.242215 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl7hj" event={"ID":"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc","Type":"ContainerDied","Data":"f9e8561e08ec1866df76a77520de09f0d2cb296c59e50127f48fe2ad0217d4fb"} Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.242242 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl7hj" event={"ID":"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc","Type":"ContainerStarted","Data":"b12997847ca460285bc1b20b78dcc282176b2b151c93435e5230fa3978e605bb"} Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.758754 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6w4bk"] Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.759683 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.762917 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.778678 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6w4bk"] Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.855497 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-utilities\") pod \"redhat-operators-6w4bk\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.855552 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-catalog-content\") pod \"redhat-operators-6w4bk\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.855611 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8kvk\" (UniqueName: \"kubernetes.io/projected/225e0c5d-e485-4157-88f6-bfb73efdac10-kube-api-access-j8kvk\") pod \"redhat-operators-6w4bk\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.956128 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-utilities\") pod \"redhat-operators-6w4bk\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.956591 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-catalog-content\") pod \"redhat-operators-6w4bk\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.956704 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fvzt6"] Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.956612 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-utilities\") pod \"redhat-operators-6w4bk\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.956823 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-catalog-content\") pod \"redhat-operators-6w4bk\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.956994 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8kvk\" (UniqueName: \"kubernetes.io/projected/225e0c5d-e485-4157-88f6-bfb73efdac10-kube-api-access-j8kvk\") pod \"redhat-operators-6w4bk\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.957829 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.961274 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 15 12:23:57 crc kubenswrapper[4719]: I1215 12:23:57.984249 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8kvk\" (UniqueName: \"kubernetes.io/projected/225e0c5d-e485-4157-88f6-bfb73efdac10-kube-api-access-j8kvk\") pod \"redhat-operators-6w4bk\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.043224 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fvzt6"] Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.078081 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.164574 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-catalog-content\") pod \"community-operators-fvzt6\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.164634 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-utilities\") pod \"community-operators-fvzt6\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.164668 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h57pt\" (UniqueName: \"kubernetes.io/projected/97f03079-f67f-4569-a9e3-6885d9c9e8cc-kube-api-access-h57pt\") pod \"community-operators-fvzt6\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.251360 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl7hj" event={"ID":"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc","Type":"ContainerStarted","Data":"19e1115aba9809b424422c33c912b56bdc15ed1d73113e0a7cfbd6290296332c"} Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.265347 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-catalog-content\") pod \"community-operators-fvzt6\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.265425 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-utilities\") pod \"community-operators-fvzt6\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.265467 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h57pt\" (UniqueName: \"kubernetes.io/projected/97f03079-f67f-4569-a9e3-6885d9c9e8cc-kube-api-access-h57pt\") pod \"community-operators-fvzt6\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.266250 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-catalog-content\") pod \"community-operators-fvzt6\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.266467 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-utilities\") pod \"community-operators-fvzt6\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.286957 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h57pt\" (UniqueName: \"kubernetes.io/projected/97f03079-f67f-4569-a9e3-6885d9c9e8cc-kube-api-access-h57pt\") pod \"community-operators-fvzt6\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.502988 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6w4bk"] Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.569958 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:23:58 crc kubenswrapper[4719]: I1215 12:23:58.973968 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fvzt6"] Dec 15 12:23:58 crc kubenswrapper[4719]: W1215 12:23:58.980387 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97f03079_f67f_4569_a9e3_6885d9c9e8cc.slice/crio-73616861ef3f4044fd85912a0850e490c31c0a87b9393c75a11f1ee73d48eb1b WatchSource:0}: Error finding container 73616861ef3f4044fd85912a0850e490c31c0a87b9393c75a11f1ee73d48eb1b: Status 404 returned error can't find the container with id 73616861ef3f4044fd85912a0850e490c31c0a87b9393c75a11f1ee73d48eb1b Dec 15 12:23:59 crc kubenswrapper[4719]: I1215 12:23:59.257327 4719 generic.go:334] "Generic (PLEG): container finished" podID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerID="8df605053a4ecfb71360d56b1d98a8a5b51cfbba2ff5203b5b77356528ba48d3" exitCode=0 Dec 15 12:23:59 crc kubenswrapper[4719]: I1215 12:23:59.257398 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fvzt6" event={"ID":"97f03079-f67f-4569-a9e3-6885d9c9e8cc","Type":"ContainerDied","Data":"8df605053a4ecfb71360d56b1d98a8a5b51cfbba2ff5203b5b77356528ba48d3"} Dec 15 12:23:59 crc kubenswrapper[4719]: I1215 12:23:59.257425 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fvzt6" event={"ID":"97f03079-f67f-4569-a9e3-6885d9c9e8cc","Type":"ContainerStarted","Data":"73616861ef3f4044fd85912a0850e490c31c0a87b9393c75a11f1ee73d48eb1b"} Dec 15 12:23:59 crc kubenswrapper[4719]: I1215 12:23:59.262564 4719 generic.go:334] "Generic (PLEG): container finished" podID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerID="985a03d56e1240db3ee814fcb6cf6beccc4ddc2e0c7f3c46b11fcc21b5115e9c" exitCode=0 Dec 15 12:23:59 crc kubenswrapper[4719]: I1215 12:23:59.262711 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w4bk" event={"ID":"225e0c5d-e485-4157-88f6-bfb73efdac10","Type":"ContainerDied","Data":"985a03d56e1240db3ee814fcb6cf6beccc4ddc2e0c7f3c46b11fcc21b5115e9c"} Dec 15 12:23:59 crc kubenswrapper[4719]: I1215 12:23:59.262792 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w4bk" event={"ID":"225e0c5d-e485-4157-88f6-bfb73efdac10","Type":"ContainerStarted","Data":"932da17b12345dcda3761fe515ade0307af0a6896827868d1bb122648faa3704"} Dec 15 12:23:59 crc kubenswrapper[4719]: I1215 12:23:59.270123 4719 generic.go:334] "Generic (PLEG): container finished" podID="44695712-4b17-480e-8c22-a392046822ca" containerID="f39de7126b137ef785f7e2f74fa5de47afd5b9a621e4f8f2edd543f4a3d1637c" exitCode=0 Dec 15 12:23:59 crc kubenswrapper[4719]: I1215 12:23:59.270194 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxvcb" event={"ID":"44695712-4b17-480e-8c22-a392046822ca","Type":"ContainerDied","Data":"f39de7126b137ef785f7e2f74fa5de47afd5b9a621e4f8f2edd543f4a3d1637c"} Dec 15 12:23:59 crc kubenswrapper[4719]: I1215 12:23:59.283658 4719 generic.go:334] "Generic (PLEG): container finished" podID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerID="19e1115aba9809b424422c33c912b56bdc15ed1d73113e0a7cfbd6290296332c" exitCode=0 Dec 15 12:23:59 crc kubenswrapper[4719]: I1215 12:23:59.283721 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl7hj" event={"ID":"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc","Type":"ContainerDied","Data":"19e1115aba9809b424422c33c912b56bdc15ed1d73113e0a7cfbd6290296332c"} Dec 15 12:24:00 crc kubenswrapper[4719]: I1215 12:24:00.289925 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxvcb" event={"ID":"44695712-4b17-480e-8c22-a392046822ca","Type":"ContainerStarted","Data":"1039bd2e6a283013c1914d81074e78a9b5ef9c36ee2daf4768271f309e87e1d3"} Dec 15 12:24:00 crc kubenswrapper[4719]: I1215 12:24:00.293777 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl7hj" event={"ID":"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc","Type":"ContainerStarted","Data":"e55f0213163fa2e4247ff82d1df64f2e99e5ecb1286b7d3e9285074c7fa62ca3"} Dec 15 12:24:00 crc kubenswrapper[4719]: I1215 12:24:00.295348 4719 generic.go:334] "Generic (PLEG): container finished" podID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerID="75ec5bf5bde70701211e415aeca48c9791f96584eebf3e0bb037276f0b8044c6" exitCode=0 Dec 15 12:24:00 crc kubenswrapper[4719]: I1215 12:24:00.295400 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fvzt6" event={"ID":"97f03079-f67f-4569-a9e3-6885d9c9e8cc","Type":"ContainerDied","Data":"75ec5bf5bde70701211e415aeca48c9791f96584eebf3e0bb037276f0b8044c6"} Dec 15 12:24:00 crc kubenswrapper[4719]: I1215 12:24:00.298652 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w4bk" event={"ID":"225e0c5d-e485-4157-88f6-bfb73efdac10","Type":"ContainerStarted","Data":"24b1f278b37b4e508f9f99ffcdaf35441ab94692e7b8d8da71098a51661d785d"} Dec 15 12:24:00 crc kubenswrapper[4719]: I1215 12:24:00.314818 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fxvcb" podStartSLOduration=2.694876898 podStartE2EDuration="5.31480108s" podCreationTimestamp="2025-12-15 12:23:55 +0000 UTC" firstStartedPulling="2025-12-15 12:23:57.240741131 +0000 UTC m=+398.183034161" lastFinishedPulling="2025-12-15 12:23:59.860665313 +0000 UTC m=+400.802958343" observedRunningTime="2025-12-15 12:24:00.312109121 +0000 UTC m=+401.254402151" watchObservedRunningTime="2025-12-15 12:24:00.31480108 +0000 UTC m=+401.257094110" Dec 15 12:24:00 crc kubenswrapper[4719]: I1215 12:24:00.353909 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xl7hj" podStartSLOduration=2.786311372 podStartE2EDuration="5.353894958s" podCreationTimestamp="2025-12-15 12:23:55 +0000 UTC" firstStartedPulling="2025-12-15 12:23:57.244128002 +0000 UTC m=+398.186421033" lastFinishedPulling="2025-12-15 12:23:59.811711589 +0000 UTC m=+400.754004619" observedRunningTime="2025-12-15 12:24:00.3527523 +0000 UTC m=+401.295045350" watchObservedRunningTime="2025-12-15 12:24:00.353894958 +0000 UTC m=+401.296187988" Dec 15 12:24:01 crc kubenswrapper[4719]: I1215 12:24:01.306368 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fvzt6" event={"ID":"97f03079-f67f-4569-a9e3-6885d9c9e8cc","Type":"ContainerStarted","Data":"7158d7d1dc94062b94b1ca9b2dcf965acf5e78932a2793d89d8de35a21c2d858"} Dec 15 12:24:01 crc kubenswrapper[4719]: I1215 12:24:01.307886 4719 generic.go:334] "Generic (PLEG): container finished" podID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerID="24b1f278b37b4e508f9f99ffcdaf35441ab94692e7b8d8da71098a51661d785d" exitCode=0 Dec 15 12:24:01 crc kubenswrapper[4719]: I1215 12:24:01.308061 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w4bk" event={"ID":"225e0c5d-e485-4157-88f6-bfb73efdac10","Type":"ContainerDied","Data":"24b1f278b37b4e508f9f99ffcdaf35441ab94692e7b8d8da71098a51661d785d"} Dec 15 12:24:01 crc kubenswrapper[4719]: I1215 12:24:01.329560 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fvzt6" podStartSLOduration=2.899333027 podStartE2EDuration="4.329542781s" podCreationTimestamp="2025-12-15 12:23:57 +0000 UTC" firstStartedPulling="2025-12-15 12:23:59.26130867 +0000 UTC m=+400.203601700" lastFinishedPulling="2025-12-15 12:24:00.691518424 +0000 UTC m=+401.633811454" observedRunningTime="2025-12-15 12:24:01.32951796 +0000 UTC m=+402.271810990" watchObservedRunningTime="2025-12-15 12:24:01.329542781 +0000 UTC m=+402.271835811" Dec 15 12:24:03 crc kubenswrapper[4719]: I1215 12:24:03.320142 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w4bk" event={"ID":"225e0c5d-e485-4157-88f6-bfb73efdac10","Type":"ContainerStarted","Data":"815a2f97be46dbde7392b009712b24479ab068c7ef6631be841cd020cbab6f2a"} Dec 15 12:24:03 crc kubenswrapper[4719]: I1215 12:24:03.336756 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6w4bk" podStartSLOduration=3.719491246 podStartE2EDuration="6.33673852s" podCreationTimestamp="2025-12-15 12:23:57 +0000 UTC" firstStartedPulling="2025-12-15 12:23:59.266885834 +0000 UTC m=+400.209178864" lastFinishedPulling="2025-12-15 12:24:01.884133108 +0000 UTC m=+402.826426138" observedRunningTime="2025-12-15 12:24:03.335567682 +0000 UTC m=+404.277860732" watchObservedRunningTime="2025-12-15 12:24:03.33673852 +0000 UTC m=+404.279031550" Dec 15 12:24:05 crc kubenswrapper[4719]: I1215 12:24:05.685604 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:24:05 crc kubenswrapper[4719]: I1215 12:24:05.686830 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:24:05 crc kubenswrapper[4719]: I1215 12:24:05.725082 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:24:05 crc kubenswrapper[4719]: I1215 12:24:05.897212 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:24:05 crc kubenswrapper[4719]: I1215 12:24:05.897552 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:24:05 crc kubenswrapper[4719]: I1215 12:24:05.932757 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:24:06 crc kubenswrapper[4719]: I1215 12:24:06.371323 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:24:06 crc kubenswrapper[4719]: I1215 12:24:06.378810 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:24:08 crc kubenswrapper[4719]: I1215 12:24:08.079409 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:24:08 crc kubenswrapper[4719]: I1215 12:24:08.079510 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:24:08 crc kubenswrapper[4719]: I1215 12:24:08.116213 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:24:08 crc kubenswrapper[4719]: I1215 12:24:08.384098 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:24:08 crc kubenswrapper[4719]: I1215 12:24:08.570991 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:24:08 crc kubenswrapper[4719]: I1215 12:24:08.571056 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:24:08 crc kubenswrapper[4719]: I1215 12:24:08.615509 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:24:09 crc kubenswrapper[4719]: I1215 12:24:09.391842 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:25:51 crc kubenswrapper[4719]: I1215 12:25:51.357412 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:25:51 crc kubenswrapper[4719]: I1215 12:25:51.358182 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:26:21 crc kubenswrapper[4719]: I1215 12:26:21.357532 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:26:21 crc kubenswrapper[4719]: I1215 12:26:21.358443 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:26:51 crc kubenswrapper[4719]: I1215 12:26:51.356499 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:26:51 crc kubenswrapper[4719]: I1215 12:26:51.357430 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:26:51 crc kubenswrapper[4719]: I1215 12:26:51.357494 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:26:51 crc kubenswrapper[4719]: I1215 12:26:51.358278 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d92d6e0171c7041e83e55a8b1507b34e41f6fa135abdaf671e5b50dd4ed3cf24"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 12:26:51 crc kubenswrapper[4719]: I1215 12:26:51.358350 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://d92d6e0171c7041e83e55a8b1507b34e41f6fa135abdaf671e5b50dd4ed3cf24" gracePeriod=600 Dec 15 12:26:51 crc kubenswrapper[4719]: I1215 12:26:51.616113 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="d92d6e0171c7041e83e55a8b1507b34e41f6fa135abdaf671e5b50dd4ed3cf24" exitCode=0 Dec 15 12:26:51 crc kubenswrapper[4719]: I1215 12:26:51.616209 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"d92d6e0171c7041e83e55a8b1507b34e41f6fa135abdaf671e5b50dd4ed3cf24"} Dec 15 12:26:51 crc kubenswrapper[4719]: I1215 12:26:51.617211 4719 scope.go:117] "RemoveContainer" containerID="38e050f4b3eb5ec1de6c0f4f194704b5cba328b77e0bccea20d90d6077185480" Dec 15 12:26:52 crc kubenswrapper[4719]: I1215 12:26:52.626361 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"1ad2c6228b0260a3d44483166c7ea0f77e7d8e32151d0daaa5541fe5e5d67cc9"} Dec 15 12:28:51 crc kubenswrapper[4719]: I1215 12:28:51.356742 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:28:51 crc kubenswrapper[4719]: I1215 12:28:51.357332 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:29:21 crc kubenswrapper[4719]: I1215 12:29:21.356572 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:29:21 crc kubenswrapper[4719]: I1215 12:29:21.357251 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.048934 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pjzlc"] Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.050331 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-pjzlc" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.053091 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.053418 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.056437 4719 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-7xnnh" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.064733 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6mllc"] Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.065378 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-6mllc" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.070017 4719 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-pr7wb" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.089248 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6mllc"] Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.097637 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-nkj5k"] Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.098365 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkj5k" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.100280 4719 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-nhjr6" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.111416 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-nkj5k"] Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.129906 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pjzlc"] Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.163644 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lrhr\" (UniqueName: \"kubernetes.io/projected/72cfd6e6-a544-4002-81ab-b57dfd2eaf94-kube-api-access-4lrhr\") pod \"cert-manager-cainjector-7f985d654d-pjzlc\" (UID: \"72cfd6e6-a544-4002-81ab-b57dfd2eaf94\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pjzlc" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.163797 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvtrp\" (UniqueName: \"kubernetes.io/projected/87e97929-9f4b-4e2e-bbc2-e9295954a8ae-kube-api-access-zvtrp\") pod \"cert-manager-5b446d88c5-6mllc\" (UID: \"87e97929-9f4b-4e2e-bbc2-e9295954a8ae\") " pod="cert-manager/cert-manager-5b446d88c5-6mllc" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.265195 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lrhr\" (UniqueName: \"kubernetes.io/projected/72cfd6e6-a544-4002-81ab-b57dfd2eaf94-kube-api-access-4lrhr\") pod \"cert-manager-cainjector-7f985d654d-pjzlc\" (UID: \"72cfd6e6-a544-4002-81ab-b57dfd2eaf94\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pjzlc" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.265334 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkqbk\" (UniqueName: \"kubernetes.io/projected/994c9412-da12-497c-8849-a5948a36e975-kube-api-access-qkqbk\") pod \"cert-manager-webhook-5655c58dd6-nkj5k\" (UID: \"994c9412-da12-497c-8849-a5948a36e975\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-nkj5k" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.265386 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvtrp\" (UniqueName: \"kubernetes.io/projected/87e97929-9f4b-4e2e-bbc2-e9295954a8ae-kube-api-access-zvtrp\") pod \"cert-manager-5b446d88c5-6mllc\" (UID: \"87e97929-9f4b-4e2e-bbc2-e9295954a8ae\") " pod="cert-manager/cert-manager-5b446d88c5-6mllc" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.288870 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvtrp\" (UniqueName: \"kubernetes.io/projected/87e97929-9f4b-4e2e-bbc2-e9295954a8ae-kube-api-access-zvtrp\") pod \"cert-manager-5b446d88c5-6mllc\" (UID: \"87e97929-9f4b-4e2e-bbc2-e9295954a8ae\") " pod="cert-manager/cert-manager-5b446d88c5-6mllc" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.294784 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lrhr\" (UniqueName: \"kubernetes.io/projected/72cfd6e6-a544-4002-81ab-b57dfd2eaf94-kube-api-access-4lrhr\") pod \"cert-manager-cainjector-7f985d654d-pjzlc\" (UID: \"72cfd6e6-a544-4002-81ab-b57dfd2eaf94\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-pjzlc" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.366567 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkqbk\" (UniqueName: \"kubernetes.io/projected/994c9412-da12-497c-8849-a5948a36e975-kube-api-access-qkqbk\") pod \"cert-manager-webhook-5655c58dd6-nkj5k\" (UID: \"994c9412-da12-497c-8849-a5948a36e975\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-nkj5k" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.368377 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-pjzlc" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.378683 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-6mllc" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.390644 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkqbk\" (UniqueName: \"kubernetes.io/projected/994c9412-da12-497c-8849-a5948a36e975-kube-api-access-qkqbk\") pod \"cert-manager-webhook-5655c58dd6-nkj5k\" (UID: \"994c9412-da12-497c-8849-a5948a36e975\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-nkj5k" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.411721 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkj5k" Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.614120 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-pjzlc"] Dec 15 12:29:34 crc kubenswrapper[4719]: W1215 12:29:34.625972 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72cfd6e6_a544_4002_81ab_b57dfd2eaf94.slice/crio-58340d8aa05cb9872780fa7e404296213742ad97b2991fe7c06ffd78f2158cab WatchSource:0}: Error finding container 58340d8aa05cb9872780fa7e404296213742ad97b2991fe7c06ffd78f2158cab: Status 404 returned error can't find the container with id 58340d8aa05cb9872780fa7e404296213742ad97b2991fe7c06ffd78f2158cab Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.628435 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.671373 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6mllc"] Dec 15 12:29:34 crc kubenswrapper[4719]: I1215 12:29:34.921479 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-nkj5k"] Dec 15 12:29:35 crc kubenswrapper[4719]: I1215 12:29:35.549240 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-pjzlc" event={"ID":"72cfd6e6-a544-4002-81ab-b57dfd2eaf94","Type":"ContainerStarted","Data":"58340d8aa05cb9872780fa7e404296213742ad97b2991fe7c06ffd78f2158cab"} Dec 15 12:29:35 crc kubenswrapper[4719]: I1215 12:29:35.550787 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkj5k" event={"ID":"994c9412-da12-497c-8849-a5948a36e975","Type":"ContainerStarted","Data":"d08145d7eb40092422cfa843e118b7e907306d09b497fc899b89ab9d093f37b3"} Dec 15 12:29:35 crc kubenswrapper[4719]: I1215 12:29:35.552315 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-6mllc" event={"ID":"87e97929-9f4b-4e2e-bbc2-e9295954a8ae","Type":"ContainerStarted","Data":"591b74e2a93e444f5b09538ee6bc51c1ef2d2d985d9b5d5ccfbf221eb180d98c"} Dec 15 12:29:39 crc kubenswrapper[4719]: I1215 12:29:39.575182 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-pjzlc" event={"ID":"72cfd6e6-a544-4002-81ab-b57dfd2eaf94","Type":"ContainerStarted","Data":"f30c9f2856d36b21647e4b463cdfdb251139fd0f08410cbb163f0fbef5bc3fab"} Dec 15 12:29:39 crc kubenswrapper[4719]: I1215 12:29:39.578057 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkj5k" event={"ID":"994c9412-da12-497c-8849-a5948a36e975","Type":"ContainerStarted","Data":"3a6643e8aac8b67051284d1ed0e042864372eceddf91421df39c14f9b2a41456"} Dec 15 12:29:39 crc kubenswrapper[4719]: I1215 12:29:39.579059 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkj5k" Dec 15 12:29:39 crc kubenswrapper[4719]: I1215 12:29:39.580160 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-6mllc" event={"ID":"87e97929-9f4b-4e2e-bbc2-e9295954a8ae","Type":"ContainerStarted","Data":"a58eb0fe2d6667e0df763f8c4dd6cc64b6279de45274681e3bf5b2b89bdf1a60"} Dec 15 12:29:39 crc kubenswrapper[4719]: I1215 12:29:39.593875 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-pjzlc" podStartSLOduration=1.883866816 podStartE2EDuration="5.593834706s" podCreationTimestamp="2025-12-15 12:29:34 +0000 UTC" firstStartedPulling="2025-12-15 12:29:34.628253141 +0000 UTC m=+735.570546171" lastFinishedPulling="2025-12-15 12:29:38.338221041 +0000 UTC m=+739.280514061" observedRunningTime="2025-12-15 12:29:39.587772381 +0000 UTC m=+740.530065421" watchObservedRunningTime="2025-12-15 12:29:39.593834706 +0000 UTC m=+740.536127756" Dec 15 12:29:39 crc kubenswrapper[4719]: I1215 12:29:39.623427 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-6mllc" podStartSLOduration=1.969221907 podStartE2EDuration="5.623409358s" podCreationTimestamp="2025-12-15 12:29:34 +0000 UTC" firstStartedPulling="2025-12-15 12:29:34.683269807 +0000 UTC m=+735.625562837" lastFinishedPulling="2025-12-15 12:29:38.337457258 +0000 UTC m=+739.279750288" observedRunningTime="2025-12-15 12:29:39.621522229 +0000 UTC m=+740.563815279" watchObservedRunningTime="2025-12-15 12:29:39.623409358 +0000 UTC m=+740.565702388" Dec 15 12:29:39 crc kubenswrapper[4719]: I1215 12:29:39.624649 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkj5k" podStartSLOduration=2.12123967 podStartE2EDuration="5.624640035s" podCreationTimestamp="2025-12-15 12:29:34 +0000 UTC" firstStartedPulling="2025-12-15 12:29:34.92289567 +0000 UTC m=+735.865188700" lastFinishedPulling="2025-12-15 12:29:38.426296035 +0000 UTC m=+739.368589065" observedRunningTime="2025-12-15 12:29:39.606446961 +0000 UTC m=+740.548739991" watchObservedRunningTime="2025-12-15 12:29:39.624640035 +0000 UTC m=+740.566933065" Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.371035 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-j8shp"] Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.372901 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovn-controller" containerID="cri-o://c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d" gracePeriod=30 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.372982 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="nbdb" containerID="cri-o://18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085" gracePeriod=30 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.373169 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="sbdb" containerID="cri-o://f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f" gracePeriod=30 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.373295 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovn-acl-logging" containerID="cri-o://ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724" gracePeriod=30 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.373286 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="kube-rbac-proxy-node" containerID="cri-o://135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da" gracePeriod=30 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.373365 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="northd" containerID="cri-o://40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae" gracePeriod=30 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.373394 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a" gracePeriod=30 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.417490 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkj5k" Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.438471 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" containerID="cri-o://407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6" gracePeriod=30 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.604544 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lcgn4_e63b3458-45ea-47c9-a9ff-7651b7dd4cca/kube-multus/2.log" Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.605148 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lcgn4_e63b3458-45ea-47c9-a9ff-7651b7dd4cca/kube-multus/1.log" Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.605192 4719 generic.go:334] "Generic (PLEG): container finished" podID="e63b3458-45ea-47c9-a9ff-7651b7dd4cca" containerID="19d7de526bd478b56554fe82bdc23648870759a6be6f40fae4d789f52a7098ee" exitCode=2 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.605255 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lcgn4" event={"ID":"e63b3458-45ea-47c9-a9ff-7651b7dd4cca","Type":"ContainerDied","Data":"19d7de526bd478b56554fe82bdc23648870759a6be6f40fae4d789f52a7098ee"} Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.605286 4719 scope.go:117] "RemoveContainer" containerID="64f9c375a5c44a4ff19da9deead918f1ce08379d6d8db7df44efe5d925055427" Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.605713 4719 scope.go:117] "RemoveContainer" containerID="19d7de526bd478b56554fe82bdc23648870759a6be6f40fae4d789f52a7098ee" Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.610580 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/3.log" Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.626247 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovn-acl-logging/0.log" Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.626901 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovn-controller/0.log" Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.627320 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a" exitCode=0 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.627342 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da" exitCode=0 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.627351 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724" exitCode=143 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.627362 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d" exitCode=143 Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.627385 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a"} Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.627413 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da"} Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.627428 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724"} Dec 15 12:29:44 crc kubenswrapper[4719]: I1215 12:29:44.627440 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d"} Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.083474 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/3.log" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.085520 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovn-acl-logging/0.log" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.086015 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovn-controller/0.log" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.086363 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146240 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gpczg"] Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146487 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146506 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146518 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovn-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146528 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovn-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146542 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146551 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146560 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="northd" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146568 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="northd" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146577 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146586 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146598 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovn-acl-logging" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146606 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovn-acl-logging" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146621 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="kube-rbac-proxy-node" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146631 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="kube-rbac-proxy-node" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146643 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="kubecfg-setup" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146651 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="kubecfg-setup" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146666 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="kube-rbac-proxy-ovn-metrics" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146674 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="kube-rbac-proxy-ovn-metrics" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146685 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="sbdb" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146693 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="sbdb" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.146702 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="nbdb" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146710 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="nbdb" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146825 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="northd" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146836 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146846 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="nbdb" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146877 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="kube-rbac-proxy-ovn-metrics" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146889 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovn-acl-logging" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146900 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146910 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146918 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146931 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovn-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146945 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="kube-rbac-proxy-node" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.146955 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="sbdb" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.147077 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.147086 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.147197 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.147304 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.147313 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerName="ovnkube-controller" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.149247 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209160 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-netd\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209465 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-env-overrides\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209482 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-openvswitch\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209510 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-netns\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209540 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-slash\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209564 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znshh\" (UniqueName: \"kubernetes.io/projected/d1afde33-4ec0-4b20-9915-1722f99997fe-kube-api-access-znshh\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209585 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1afde33-4ec0-4b20-9915-1722f99997fe-ovn-node-metrics-cert\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209600 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-log-socket\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209613 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-bin\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209627 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-var-lib-cni-networks-ovn-kubernetes\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209642 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-ovn-kubernetes\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209660 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-etc-openvswitch\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209680 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-script-lib\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209697 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-systemd\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209721 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-kubelet\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209770 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-systemd-units\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209786 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-ovn\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209805 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-var-lib-openvswitch\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209831 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-node-log\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209878 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-config\") pod \"d1afde33-4ec0-4b20-9915-1722f99997fe\" (UID: \"d1afde33-4ec0-4b20-9915-1722f99997fe\") " Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209993 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr7cn\" (UniqueName: \"kubernetes.io/projected/12064236-8548-46aa-82f5-51e91d394eda-kube-api-access-xr7cn\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210018 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210040 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-run-ovn-kubernetes\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210058 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/12064236-8548-46aa-82f5-51e91d394eda-ovn-node-metrics-cert\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210078 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-run-netns\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210098 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/12064236-8548-46aa-82f5-51e91d394eda-ovnkube-script-lib\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210114 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/12064236-8548-46aa-82f5-51e91d394eda-env-overrides\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210130 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-run-openvswitch\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210149 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-cni-netd\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210163 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-run-systemd\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210178 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-run-ovn\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210197 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-slash\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210213 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-systemd-units\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210227 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-node-log\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210247 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-kubelet\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210263 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/12064236-8548-46aa-82f5-51e91d394eda-ovnkube-config\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210281 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-var-lib-openvswitch\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210299 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-etc-openvswitch\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210316 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-log-socket\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210334 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-cni-bin\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.209301 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210667 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210742 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210762 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210911 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210945 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-slash" (OuterVolumeSpecName: "host-slash") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210978 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.210998 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-log-socket" (OuterVolumeSpecName: "log-socket") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.211019 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.211001 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.211051 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.211077 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.211100 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-node-log" (OuterVolumeSpecName: "node-log") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.211124 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.211443 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.211497 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.211826 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.217957 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1afde33-4ec0-4b20-9915-1722f99997fe-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.218071 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1afde33-4ec0-4b20-9915-1722f99997fe-kube-api-access-znshh" (OuterVolumeSpecName: "kube-api-access-znshh") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "kube-api-access-znshh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.227627 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "d1afde33-4ec0-4b20-9915-1722f99997fe" (UID: "d1afde33-4ec0-4b20-9915-1722f99997fe"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311539 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-run-ovn-kubernetes\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311619 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/12064236-8548-46aa-82f5-51e91d394eda-ovn-node-metrics-cert\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311642 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-run-netns\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311650 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-run-ovn-kubernetes\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311665 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/12064236-8548-46aa-82f5-51e91d394eda-ovnkube-script-lib\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311742 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-run-netns\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311767 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-run-openvswitch\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311788 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/12064236-8548-46aa-82f5-51e91d394eda-env-overrides\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311812 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-run-openvswitch\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311826 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-run-systemd\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311844 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-cni-netd\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311960 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-run-ovn\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.311990 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-slash\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312017 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-systemd-units\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312035 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-node-log\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312074 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-kubelet\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312102 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/12064236-8548-46aa-82f5-51e91d394eda-ovnkube-config\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312108 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-cni-netd\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312142 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-var-lib-openvswitch\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312182 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-etc-openvswitch\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312178 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-run-systemd\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312210 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-log-socket\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312235 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-node-log\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312241 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-cni-bin\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312263 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-run-ovn\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312274 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr7cn\" (UniqueName: \"kubernetes.io/projected/12064236-8548-46aa-82f5-51e91d394eda-kube-api-access-xr7cn\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312286 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-slash\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312308 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-systemd-units\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312314 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312334 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-etc-openvswitch\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312364 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/12064236-8548-46aa-82f5-51e91d394eda-env-overrides\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312379 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/12064236-8548-46aa-82f5-51e91d394eda-ovnkube-script-lib\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312393 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-log-socket\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312397 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-cni-bin\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312402 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-kubelet\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312548 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312576 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12064236-8548-46aa-82f5-51e91d394eda-var-lib-openvswitch\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312678 4719 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312691 4719 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312700 4719 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312708 4719 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312718 4719 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-slash\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312726 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znshh\" (UniqueName: \"kubernetes.io/projected/d1afde33-4ec0-4b20-9915-1722f99997fe-kube-api-access-znshh\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312735 4719 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d1afde33-4ec0-4b20-9915-1722f99997fe-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312770 4719 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-log-socket\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312787 4719 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312802 4719 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312816 4719 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312830 4719 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312842 4719 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312883 4719 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312896 4719 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312909 4719 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312920 4719 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312932 4719 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312942 4719 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d1afde33-4ec0-4b20-9915-1722f99997fe-node-log\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312953 4719 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d1afde33-4ec0-4b20-9915-1722f99997fe-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.312999 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/12064236-8548-46aa-82f5-51e91d394eda-ovnkube-config\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.316385 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/12064236-8548-46aa-82f5-51e91d394eda-ovn-node-metrics-cert\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.329570 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr7cn\" (UniqueName: \"kubernetes.io/projected/12064236-8548-46aa-82f5-51e91d394eda-kube-api-access-xr7cn\") pod \"ovnkube-node-gpczg\" (UID: \"12064236-8548-46aa-82f5-51e91d394eda\") " pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.464199 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:45 crc kubenswrapper[4719]: W1215 12:29:45.491489 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12064236_8548_46aa_82f5_51e91d394eda.slice/crio-a40391d3d7d6c93ddb05fc4d08d18a3e0c94098a4fd3010684c76e1b2a7bb6b7 WatchSource:0}: Error finding container a40391d3d7d6c93ddb05fc4d08d18a3e0c94098a4fd3010684c76e1b2a7bb6b7: Status 404 returned error can't find the container with id a40391d3d7d6c93ddb05fc4d08d18a3e0c94098a4fd3010684c76e1b2a7bb6b7 Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.634272 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" event={"ID":"12064236-8548-46aa-82f5-51e91d394eda","Type":"ContainerStarted","Data":"a40391d3d7d6c93ddb05fc4d08d18a3e0c94098a4fd3010684c76e1b2a7bb6b7"} Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.636243 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lcgn4_e63b3458-45ea-47c9-a9ff-7651b7dd4cca/kube-multus/2.log" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.636410 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lcgn4" event={"ID":"e63b3458-45ea-47c9-a9ff-7651b7dd4cca","Type":"ContainerStarted","Data":"8733d2d862f1fe6d8a5576f24e7164a800dddadde73abeb9cbb0a2126da572a2"} Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.639660 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovnkube-controller/3.log" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.645207 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovn-acl-logging/0.log" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.645913 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-j8shp_d1afde33-4ec0-4b20-9915-1722f99997fe/ovn-controller/0.log" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.646629 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6" exitCode=0 Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.646661 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f" exitCode=0 Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.646671 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085" exitCode=0 Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.646681 4719 generic.go:334] "Generic (PLEG): container finished" podID="d1afde33-4ec0-4b20-9915-1722f99997fe" containerID="40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae" exitCode=0 Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.646723 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6"} Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.646784 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f"} Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.646799 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085"} Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.646815 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae"} Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.646826 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" event={"ID":"d1afde33-4ec0-4b20-9915-1722f99997fe","Type":"ContainerDied","Data":"9f3b314ffd9a827cbb7b0808c98f3c094e1b82cac64eb7e739dada30ad41321b"} Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.646845 4719 scope.go:117] "RemoveContainer" containerID="407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.647047 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-j8shp" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.726342 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-j8shp"] Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.726375 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.729379 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-j8shp"] Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.743585 4719 scope.go:117] "RemoveContainer" containerID="f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.762299 4719 scope.go:117] "RemoveContainer" containerID="18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.775706 4719 scope.go:117] "RemoveContainer" containerID="40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.788724 4719 scope.go:117] "RemoveContainer" containerID="083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.802944 4719 scope.go:117] "RemoveContainer" containerID="135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.816514 4719 scope.go:117] "RemoveContainer" containerID="ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.828735 4719 scope.go:117] "RemoveContainer" containerID="c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.844555 4719 scope.go:117] "RemoveContainer" containerID="23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.861327 4719 scope.go:117] "RemoveContainer" containerID="407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.861941 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6\": container with ID starting with 407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6 not found: ID does not exist" containerID="407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.861979 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6"} err="failed to get container status \"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6\": rpc error: code = NotFound desc = could not find container \"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6\": container with ID starting with 407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.862003 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.862652 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\": container with ID starting with 44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1 not found: ID does not exist" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.862674 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1"} err="failed to get container status \"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\": rpc error: code = NotFound desc = could not find container \"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\": container with ID starting with 44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.862688 4719 scope.go:117] "RemoveContainer" containerID="f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.863520 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\": container with ID starting with f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f not found: ID does not exist" containerID="f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.863552 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f"} err="failed to get container status \"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\": rpc error: code = NotFound desc = could not find container \"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\": container with ID starting with f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.863566 4719 scope.go:117] "RemoveContainer" containerID="18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.864105 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\": container with ID starting with 18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085 not found: ID does not exist" containerID="18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.864128 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085"} err="failed to get container status \"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\": rpc error: code = NotFound desc = could not find container \"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\": container with ID starting with 18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.864159 4719 scope.go:117] "RemoveContainer" containerID="40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.864363 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\": container with ID starting with 40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae not found: ID does not exist" containerID="40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.864400 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae"} err="failed to get container status \"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\": rpc error: code = NotFound desc = could not find container \"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\": container with ID starting with 40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.864413 4719 scope.go:117] "RemoveContainer" containerID="083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.864787 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\": container with ID starting with 083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a not found: ID does not exist" containerID="083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.864805 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a"} err="failed to get container status \"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\": rpc error: code = NotFound desc = could not find container \"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\": container with ID starting with 083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.864833 4719 scope.go:117] "RemoveContainer" containerID="135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.865135 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\": container with ID starting with 135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da not found: ID does not exist" containerID="135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.865153 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da"} err="failed to get container status \"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\": rpc error: code = NotFound desc = could not find container \"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\": container with ID starting with 135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.865167 4719 scope.go:117] "RemoveContainer" containerID="ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.865893 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\": container with ID starting with ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724 not found: ID does not exist" containerID="ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.865914 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724"} err="failed to get container status \"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\": rpc error: code = NotFound desc = could not find container \"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\": container with ID starting with ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.865926 4719 scope.go:117] "RemoveContainer" containerID="c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.866343 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\": container with ID starting with c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d not found: ID does not exist" containerID="c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.866379 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d"} err="failed to get container status \"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\": rpc error: code = NotFound desc = could not find container \"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\": container with ID starting with c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.866392 4719 scope.go:117] "RemoveContainer" containerID="23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5" Dec 15 12:29:45 crc kubenswrapper[4719]: E1215 12:29:45.866768 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\": container with ID starting with 23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5 not found: ID does not exist" containerID="23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.866789 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5"} err="failed to get container status \"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\": rpc error: code = NotFound desc = could not find container \"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\": container with ID starting with 23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.866817 4719 scope.go:117] "RemoveContainer" containerID="407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.867260 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6"} err="failed to get container status \"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6\": rpc error: code = NotFound desc = could not find container \"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6\": container with ID starting with 407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.867277 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.867603 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1"} err="failed to get container status \"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\": rpc error: code = NotFound desc = could not find container \"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\": container with ID starting with 44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.867619 4719 scope.go:117] "RemoveContainer" containerID="f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.867847 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f"} err="failed to get container status \"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\": rpc error: code = NotFound desc = could not find container \"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\": container with ID starting with f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.867893 4719 scope.go:117] "RemoveContainer" containerID="18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.868409 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085"} err="failed to get container status \"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\": rpc error: code = NotFound desc = could not find container \"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\": container with ID starting with 18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.868427 4719 scope.go:117] "RemoveContainer" containerID="40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.868761 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae"} err="failed to get container status \"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\": rpc error: code = NotFound desc = could not find container \"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\": container with ID starting with 40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.868780 4719 scope.go:117] "RemoveContainer" containerID="083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.869040 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a"} err="failed to get container status \"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\": rpc error: code = NotFound desc = could not find container \"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\": container with ID starting with 083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.869062 4719 scope.go:117] "RemoveContainer" containerID="135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.869325 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da"} err="failed to get container status \"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\": rpc error: code = NotFound desc = could not find container \"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\": container with ID starting with 135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.869342 4719 scope.go:117] "RemoveContainer" containerID="ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.869636 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724"} err="failed to get container status \"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\": rpc error: code = NotFound desc = could not find container \"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\": container with ID starting with ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.869653 4719 scope.go:117] "RemoveContainer" containerID="c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.869954 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d"} err="failed to get container status \"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\": rpc error: code = NotFound desc = could not find container \"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\": container with ID starting with c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.869970 4719 scope.go:117] "RemoveContainer" containerID="23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.870274 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5"} err="failed to get container status \"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\": rpc error: code = NotFound desc = could not find container \"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\": container with ID starting with 23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.870292 4719 scope.go:117] "RemoveContainer" containerID="407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.871196 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6"} err="failed to get container status \"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6\": rpc error: code = NotFound desc = could not find container \"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6\": container with ID starting with 407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.871271 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.871610 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1"} err="failed to get container status \"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\": rpc error: code = NotFound desc = could not find container \"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\": container with ID starting with 44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.871679 4719 scope.go:117] "RemoveContainer" containerID="f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.871981 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f"} err="failed to get container status \"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\": rpc error: code = NotFound desc = could not find container \"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\": container with ID starting with f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.872006 4719 scope.go:117] "RemoveContainer" containerID="18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.872334 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085"} err="failed to get container status \"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\": rpc error: code = NotFound desc = could not find container \"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\": container with ID starting with 18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.872387 4719 scope.go:117] "RemoveContainer" containerID="40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.872793 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae"} err="failed to get container status \"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\": rpc error: code = NotFound desc = could not find container \"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\": container with ID starting with 40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.872827 4719 scope.go:117] "RemoveContainer" containerID="083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.873103 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a"} err="failed to get container status \"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\": rpc error: code = NotFound desc = could not find container \"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\": container with ID starting with 083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.873154 4719 scope.go:117] "RemoveContainer" containerID="135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.873442 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da"} err="failed to get container status \"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\": rpc error: code = NotFound desc = could not find container \"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\": container with ID starting with 135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.873500 4719 scope.go:117] "RemoveContainer" containerID="ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.873746 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724"} err="failed to get container status \"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\": rpc error: code = NotFound desc = could not find container \"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\": container with ID starting with ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.873767 4719 scope.go:117] "RemoveContainer" containerID="c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.873958 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d"} err="failed to get container status \"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\": rpc error: code = NotFound desc = could not find container \"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\": container with ID starting with c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.873990 4719 scope.go:117] "RemoveContainer" containerID="23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.874232 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5"} err="failed to get container status \"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\": rpc error: code = NotFound desc = could not find container \"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\": container with ID starting with 23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.874255 4719 scope.go:117] "RemoveContainer" containerID="407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.874709 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6"} err="failed to get container status \"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6\": rpc error: code = NotFound desc = could not find container \"407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6\": container with ID starting with 407ebcc2a76cdcc20be3519a2ae1488b0ebc1e9363ffe9f4306c9ac5e6d2f5f6 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.874733 4719 scope.go:117] "RemoveContainer" containerID="44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.875056 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1"} err="failed to get container status \"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\": rpc error: code = NotFound desc = could not find container \"44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1\": container with ID starting with 44a2cd2d3d61a7bec66a3079851e0c47ba135946b5a69921526d2e99fcd39cd1 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.875077 4719 scope.go:117] "RemoveContainer" containerID="f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.875359 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f"} err="failed to get container status \"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\": rpc error: code = NotFound desc = could not find container \"f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f\": container with ID starting with f9c73e1f0a93451184e7e138c68f80a662f347a4b301f2bd1a45e9997941b55f not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.875378 4719 scope.go:117] "RemoveContainer" containerID="18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.875603 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085"} err="failed to get container status \"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\": rpc error: code = NotFound desc = could not find container \"18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085\": container with ID starting with 18f5976524b29ccab8b54759a4a8657ac2dfaeac3bf006eadf097b317cd5e085 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.875655 4719 scope.go:117] "RemoveContainer" containerID="40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.875954 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae"} err="failed to get container status \"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\": rpc error: code = NotFound desc = could not find container \"40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae\": container with ID starting with 40d467dfa65871fa0360bf39dfc05c9a7d66d345291a1e7bbca5a4351bb07bae not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.875981 4719 scope.go:117] "RemoveContainer" containerID="083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.876213 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a"} err="failed to get container status \"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\": rpc error: code = NotFound desc = could not find container \"083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a\": container with ID starting with 083adb3f5a825e827d5806e26b075409331d69b6e013cf9c9c0ed73200ee3b9a not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.876239 4719 scope.go:117] "RemoveContainer" containerID="135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.876469 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da"} err="failed to get container status \"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\": rpc error: code = NotFound desc = could not find container \"135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da\": container with ID starting with 135956d8da118d106f2dee7c0bdade0a9e2d1aa5c561fb5f5fefccdd818441da not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.876491 4719 scope.go:117] "RemoveContainer" containerID="ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.876718 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724"} err="failed to get container status \"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\": rpc error: code = NotFound desc = could not find container \"ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724\": container with ID starting with ba6b3f18af6d4b73984c11d9e27ceb2c72310f9a29ef8387b83ddf1aeacd5724 not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.876740 4719 scope.go:117] "RemoveContainer" containerID="c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.877077 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d"} err="failed to get container status \"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\": rpc error: code = NotFound desc = could not find container \"c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d\": container with ID starting with c7c7ec436e9f62ac84881340892f1fe81e68beec131e29c14d5b5a578d19a59d not found: ID does not exist" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.877104 4719 scope.go:117] "RemoveContainer" containerID="23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5" Dec 15 12:29:45 crc kubenswrapper[4719]: I1215 12:29:45.877377 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5"} err="failed to get container status \"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\": rpc error: code = NotFound desc = could not find container \"23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5\": container with ID starting with 23c5179890bb2822bb6e4ead93d3e1b25ead9d9826b873286d357e9ebb9611d5 not found: ID does not exist" Dec 15 12:29:46 crc kubenswrapper[4719]: I1215 12:29:46.651952 4719 generic.go:334] "Generic (PLEG): container finished" podID="12064236-8548-46aa-82f5-51e91d394eda" containerID="2e323c3cf32e2863a757c4e7fdd8ee1520824fed406945609273779668a5d89e" exitCode=0 Dec 15 12:29:46 crc kubenswrapper[4719]: I1215 12:29:46.652011 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" event={"ID":"12064236-8548-46aa-82f5-51e91d394eda","Type":"ContainerDied","Data":"2e323c3cf32e2863a757c4e7fdd8ee1520824fed406945609273779668a5d89e"} Dec 15 12:29:47 crc kubenswrapper[4719]: I1215 12:29:47.482040 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1afde33-4ec0-4b20-9915-1722f99997fe" path="/var/lib/kubelet/pods/d1afde33-4ec0-4b20-9915-1722f99997fe/volumes" Dec 15 12:29:47 crc kubenswrapper[4719]: I1215 12:29:47.661850 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" event={"ID":"12064236-8548-46aa-82f5-51e91d394eda","Type":"ContainerStarted","Data":"4bbef80c31feb6e194e934f83113faff9ba507d7179153e5702938a85cf28d0a"} Dec 15 12:29:47 crc kubenswrapper[4719]: I1215 12:29:47.662711 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" event={"ID":"12064236-8548-46aa-82f5-51e91d394eda","Type":"ContainerStarted","Data":"4872735a6818d2274c28eb5cbac0740fae227749e68a4afce01c1155d814de64"} Dec 15 12:29:47 crc kubenswrapper[4719]: I1215 12:29:47.662827 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" event={"ID":"12064236-8548-46aa-82f5-51e91d394eda","Type":"ContainerStarted","Data":"caf7c23c1b9f1a24539d0c7add885aa55839f5b3cd101eab107d157e0321985a"} Dec 15 12:29:47 crc kubenswrapper[4719]: I1215 12:29:47.662919 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" event={"ID":"12064236-8548-46aa-82f5-51e91d394eda","Type":"ContainerStarted","Data":"8b554f77667363617d1f4c092c0ded997e0734a2c63645f51555917b8a23211b"} Dec 15 12:29:47 crc kubenswrapper[4719]: I1215 12:29:47.662974 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" event={"ID":"12064236-8548-46aa-82f5-51e91d394eda","Type":"ContainerStarted","Data":"0f2e653afb63965555f2c2887e0f7010b1cd5391f89dafe2b1cf9055d64b7c77"} Dec 15 12:29:47 crc kubenswrapper[4719]: I1215 12:29:47.663028 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" event={"ID":"12064236-8548-46aa-82f5-51e91d394eda","Type":"ContainerStarted","Data":"a9945d4ab3d0c9846bc85e29310d7636db42e74713798952714ca9407dd06bb1"} Dec 15 12:29:49 crc kubenswrapper[4719]: I1215 12:29:49.678603 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" event={"ID":"12064236-8548-46aa-82f5-51e91d394eda","Type":"ContainerStarted","Data":"744ea64a7c1f1481f4724333b3bbdf2ba0b4a65ad8961f21e419f9bf92aaf728"} Dec 15 12:29:51 crc kubenswrapper[4719]: I1215 12:29:51.356578 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:29:51 crc kubenswrapper[4719]: I1215 12:29:51.357156 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:29:51 crc kubenswrapper[4719]: I1215 12:29:51.357233 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:29:51 crc kubenswrapper[4719]: I1215 12:29:51.357877 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1ad2c6228b0260a3d44483166c7ea0f77e7d8e32151d0daaa5541fe5e5d67cc9"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 12:29:51 crc kubenswrapper[4719]: I1215 12:29:51.357949 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://1ad2c6228b0260a3d44483166c7ea0f77e7d8e32151d0daaa5541fe5e5d67cc9" gracePeriod=600 Dec 15 12:29:52 crc kubenswrapper[4719]: I1215 12:29:52.699006 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="1ad2c6228b0260a3d44483166c7ea0f77e7d8e32151d0daaa5541fe5e5d67cc9" exitCode=0 Dec 15 12:29:52 crc kubenswrapper[4719]: I1215 12:29:52.699020 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"1ad2c6228b0260a3d44483166c7ea0f77e7d8e32151d0daaa5541fe5e5d67cc9"} Dec 15 12:29:52 crc kubenswrapper[4719]: I1215 12:29:52.700763 4719 scope.go:117] "RemoveContainer" containerID="d92d6e0171c7041e83e55a8b1507b34e41f6fa135abdaf671e5b50dd4ed3cf24" Dec 15 12:29:53 crc kubenswrapper[4719]: I1215 12:29:53.710576 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" event={"ID":"12064236-8548-46aa-82f5-51e91d394eda","Type":"ContainerStarted","Data":"1472cc6eb0c7c009c47a65d1624a9135d517d21cd8c06cc3e575c03990186212"} Dec 15 12:29:53 crc kubenswrapper[4719]: I1215 12:29:53.710960 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:53 crc kubenswrapper[4719]: I1215 12:29:53.710978 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:53 crc kubenswrapper[4719]: I1215 12:29:53.710990 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:53 crc kubenswrapper[4719]: I1215 12:29:53.744824 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:53 crc kubenswrapper[4719]: I1215 12:29:53.745340 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:29:53 crc kubenswrapper[4719]: I1215 12:29:53.749668 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" podStartSLOduration=8.749655802 podStartE2EDuration="8.749655802s" podCreationTimestamp="2025-12-15 12:29:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:29:53.748830087 +0000 UTC m=+754.691123107" watchObservedRunningTime="2025-12-15 12:29:53.749655802 +0000 UTC m=+754.691948832" Dec 15 12:29:54 crc kubenswrapper[4719]: I1215 12:29:54.717534 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"2d86014a7f0b55979e42be548b06831288da0c2f9ca42323013a65f25c4ad994"} Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.158019 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb"] Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.159140 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.163984 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.164338 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.167998 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb"] Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.301979 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-secret-volume\") pod \"collect-profiles-29430030-vc5hb\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.302056 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-config-volume\") pod \"collect-profiles-29430030-vc5hb\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.302306 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d56xt\" (UniqueName: \"kubernetes.io/projected/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-kube-api-access-d56xt\") pod \"collect-profiles-29430030-vc5hb\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.403395 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d56xt\" (UniqueName: \"kubernetes.io/projected/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-kube-api-access-d56xt\") pod \"collect-profiles-29430030-vc5hb\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.403498 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-secret-volume\") pod \"collect-profiles-29430030-vc5hb\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.403539 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-config-volume\") pod \"collect-profiles-29430030-vc5hb\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.404608 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-config-volume\") pod \"collect-profiles-29430030-vc5hb\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.416753 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-secret-volume\") pod \"collect-profiles-29430030-vc5hb\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.423518 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d56xt\" (UniqueName: \"kubernetes.io/projected/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-kube-api-access-d56xt\") pod \"collect-profiles-29430030-vc5hb\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.484659 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:00 crc kubenswrapper[4719]: I1215 12:30:00.877766 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb"] Dec 15 12:30:01 crc kubenswrapper[4719]: I1215 12:30:01.756776 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" event={"ID":"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5","Type":"ContainerStarted","Data":"97aaa509878268d2bb03c24a14b7861cbc39f40b96f5f043bf9e693599cba0c5"} Dec 15 12:30:01 crc kubenswrapper[4719]: I1215 12:30:01.757184 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" event={"ID":"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5","Type":"ContainerStarted","Data":"f1eaad2567e3fd9f8431924ce6485c5b7a2ed865bbc9c9e0448a4cf9945bdb8a"} Dec 15 12:30:02 crc kubenswrapper[4719]: I1215 12:30:02.764400 4719 generic.go:334] "Generic (PLEG): container finished" podID="e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5" containerID="97aaa509878268d2bb03c24a14b7861cbc39f40b96f5f043bf9e693599cba0c5" exitCode=0 Dec 15 12:30:02 crc kubenswrapper[4719]: I1215 12:30:02.764437 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" event={"ID":"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5","Type":"ContainerDied","Data":"97aaa509878268d2bb03c24a14b7861cbc39f40b96f5f043bf9e693599cba0c5"} Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.068555 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.145268 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-config-volume\") pod \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.145410 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d56xt\" (UniqueName: \"kubernetes.io/projected/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-kube-api-access-d56xt\") pod \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.146130 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-secret-volume\") pod \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\" (UID: \"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5\") " Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.146134 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-config-volume" (OuterVolumeSpecName: "config-volume") pod "e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5" (UID: "e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.146779 4719 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.152127 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-kube-api-access-d56xt" (OuterVolumeSpecName: "kube-api-access-d56xt") pod "e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5" (UID: "e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5"). InnerVolumeSpecName "kube-api-access-d56xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.152488 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5" (UID: "e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.247995 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d56xt\" (UniqueName: \"kubernetes.io/projected/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-kube-api-access-d56xt\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.248051 4719 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.777669 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" event={"ID":"e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5","Type":"ContainerDied","Data":"f1eaad2567e3fd9f8431924ce6485c5b7a2ed865bbc9c9e0448a4cf9945bdb8a"} Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.778113 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1eaad2567e3fd9f8431924ce6485c5b7a2ed865bbc9c9e0448a4cf9945bdb8a" Dec 15 12:30:04 crc kubenswrapper[4719]: I1215 12:30:04.777747 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb" Dec 15 12:30:10 crc kubenswrapper[4719]: I1215 12:30:10.783647 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nxt4c"] Dec 15 12:30:10 crc kubenswrapper[4719]: E1215 12:30:10.784197 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5" containerName="collect-profiles" Dec 15 12:30:10 crc kubenswrapper[4719]: I1215 12:30:10.784210 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5" containerName="collect-profiles" Dec 15 12:30:10 crc kubenswrapper[4719]: I1215 12:30:10.784328 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5" containerName="collect-profiles" Dec 15 12:30:10 crc kubenswrapper[4719]: I1215 12:30:10.785073 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:10 crc kubenswrapper[4719]: I1215 12:30:10.810580 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nxt4c"] Dec 15 12:30:10 crc kubenswrapper[4719]: I1215 12:30:10.929153 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db5mj\" (UniqueName: \"kubernetes.io/projected/73cbed55-0dac-4738-b4e2-0d863c20570f-kube-api-access-db5mj\") pod \"community-operators-nxt4c\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:10 crc kubenswrapper[4719]: I1215 12:30:10.929242 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-catalog-content\") pod \"community-operators-nxt4c\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:10 crc kubenswrapper[4719]: I1215 12:30:10.929314 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-utilities\") pod \"community-operators-nxt4c\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.030500 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-utilities\") pod \"community-operators-nxt4c\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.030772 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db5mj\" (UniqueName: \"kubernetes.io/projected/73cbed55-0dac-4738-b4e2-0d863c20570f-kube-api-access-db5mj\") pod \"community-operators-nxt4c\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.030883 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-catalog-content\") pod \"community-operators-nxt4c\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.031196 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-utilities\") pod \"community-operators-nxt4c\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.031286 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-catalog-content\") pod \"community-operators-nxt4c\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.055396 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db5mj\" (UniqueName: \"kubernetes.io/projected/73cbed55-0dac-4738-b4e2-0d863c20570f-kube-api-access-db5mj\") pod \"community-operators-nxt4c\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.115319 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.367947 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nxt4c"] Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.814491 4719 generic.go:334] "Generic (PLEG): container finished" podID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerID="67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6" exitCode=0 Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.814527 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxt4c" event={"ID":"73cbed55-0dac-4738-b4e2-0d863c20570f","Type":"ContainerDied","Data":"67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6"} Dec 15 12:30:11 crc kubenswrapper[4719]: I1215 12:30:11.815154 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxt4c" event={"ID":"73cbed55-0dac-4738-b4e2-0d863c20570f","Type":"ContainerStarted","Data":"edffa6e426dff1abef5bb791bf2e4bc4019c2a6e1df2b09bf3d4d962599535ee"} Dec 15 12:30:12 crc kubenswrapper[4719]: I1215 12:30:12.821110 4719 generic.go:334] "Generic (PLEG): container finished" podID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerID="6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62" exitCode=0 Dec 15 12:30:12 crc kubenswrapper[4719]: I1215 12:30:12.821471 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxt4c" event={"ID":"73cbed55-0dac-4738-b4e2-0d863c20570f","Type":"ContainerDied","Data":"6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62"} Dec 15 12:30:13 crc kubenswrapper[4719]: I1215 12:30:13.830081 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxt4c" event={"ID":"73cbed55-0dac-4738-b4e2-0d863c20570f","Type":"ContainerStarted","Data":"9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60"} Dec 15 12:30:13 crc kubenswrapper[4719]: I1215 12:30:13.852971 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nxt4c" podStartSLOduration=2.36788994 podStartE2EDuration="3.852953277s" podCreationTimestamp="2025-12-15 12:30:10 +0000 UTC" firstStartedPulling="2025-12-15 12:30:11.816210847 +0000 UTC m=+772.758503887" lastFinishedPulling="2025-12-15 12:30:13.301274194 +0000 UTC m=+774.243567224" observedRunningTime="2025-12-15 12:30:13.848318206 +0000 UTC m=+774.790611236" watchObservedRunningTime="2025-12-15 12:30:13.852953277 +0000 UTC m=+774.795246317" Dec 15 12:30:15 crc kubenswrapper[4719]: I1215 12:30:15.493294 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gpczg" Dec 15 12:30:21 crc kubenswrapper[4719]: I1215 12:30:21.116028 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:21 crc kubenswrapper[4719]: I1215 12:30:21.116605 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:21 crc kubenswrapper[4719]: I1215 12:30:21.155244 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:21 crc kubenswrapper[4719]: I1215 12:30:21.912886 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:21 crc kubenswrapper[4719]: I1215 12:30:21.950459 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nxt4c"] Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.256254 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xl7hj"] Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.256567 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xl7hj" podUID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerName="registry-server" containerID="cri-o://e55f0213163fa2e4247ff82d1df64f2e99e5ecb1286b7d3e9285074c7fa62ca3" gracePeriod=30 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.266515 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fvzt6"] Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.266911 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fvzt6" podUID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerName="registry-server" containerID="cri-o://7158d7d1dc94062b94b1ca9b2dcf965acf5e78932a2793d89d8de35a21c2d858" gracePeriod=30 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.273927 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ds7r2"] Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.274313 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" podUID="5de2277b-38a8-4be5-b4f5-b7516197192c" containerName="marketplace-operator" containerID="cri-o://3a6d10cfe2ade7410f3ac7ea893a64ff7cb491f5ec4e015f333f7eedd228f538" gracePeriod=30 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.288647 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxvcb"] Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.288985 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fxvcb" podUID="44695712-4b17-480e-8c22-a392046822ca" containerName="registry-server" containerID="cri-o://1039bd2e6a283013c1914d81074e78a9b5ef9c36ee2daf4768271f309e87e1d3" gracePeriod=30 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.297891 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6w4bk"] Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.298304 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6w4bk" podUID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerName="registry-server" containerID="cri-o://815a2f97be46dbde7392b009712b24479ab068c7ef6631be841cd020cbab6f2a" gracePeriod=30 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.302107 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kxrn8"] Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.302908 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.323874 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kxrn8"] Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.378391 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e804e02c-d21b-4881-9bc8-2de04a0928f9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kxrn8\" (UID: \"e804e02c-d21b-4881-9bc8-2de04a0928f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.378504 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e804e02c-d21b-4881-9bc8-2de04a0928f9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kxrn8\" (UID: \"e804e02c-d21b-4881-9bc8-2de04a0928f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.378663 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t82nk\" (UniqueName: \"kubernetes.io/projected/e804e02c-d21b-4881-9bc8-2de04a0928f9-kube-api-access-t82nk\") pod \"marketplace-operator-79b997595-kxrn8\" (UID: \"e804e02c-d21b-4881-9bc8-2de04a0928f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.479561 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e804e02c-d21b-4881-9bc8-2de04a0928f9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kxrn8\" (UID: \"e804e02c-d21b-4881-9bc8-2de04a0928f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.479632 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e804e02c-d21b-4881-9bc8-2de04a0928f9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kxrn8\" (UID: \"e804e02c-d21b-4881-9bc8-2de04a0928f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.479696 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t82nk\" (UniqueName: \"kubernetes.io/projected/e804e02c-d21b-4881-9bc8-2de04a0928f9-kube-api-access-t82nk\") pod \"marketplace-operator-79b997595-kxrn8\" (UID: \"e804e02c-d21b-4881-9bc8-2de04a0928f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.481446 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e804e02c-d21b-4881-9bc8-2de04a0928f9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kxrn8\" (UID: \"e804e02c-d21b-4881-9bc8-2de04a0928f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.487709 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e804e02c-d21b-4881-9bc8-2de04a0928f9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kxrn8\" (UID: \"e804e02c-d21b-4881-9bc8-2de04a0928f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.499062 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t82nk\" (UniqueName: \"kubernetes.io/projected/e804e02c-d21b-4881-9bc8-2de04a0928f9-kube-api-access-t82nk\") pod \"marketplace-operator-79b997595-kxrn8\" (UID: \"e804e02c-d21b-4881-9bc8-2de04a0928f9\") " pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.516371 4719 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ds7r2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" start-of-body= Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.516742 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" podUID="5de2277b-38a8-4be5-b4f5-b7516197192c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.625706 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.796893 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kxrn8"] Dec 15 12:30:23 crc kubenswrapper[4719]: W1215 12:30:23.843959 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode804e02c_d21b_4881_9bc8_2de04a0928f9.slice/crio-1ef928bc1d41a830ececb59b97767d7da4de79893999255b84ccbbdbc3ea7809 WatchSource:0}: Error finding container 1ef928bc1d41a830ececb59b97767d7da4de79893999255b84ccbbdbc3ea7809: Status 404 returned error can't find the container with id 1ef928bc1d41a830ececb59b97767d7da4de79893999255b84ccbbdbc3ea7809 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.888829 4719 generic.go:334] "Generic (PLEG): container finished" podID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerID="815a2f97be46dbde7392b009712b24479ab068c7ef6631be841cd020cbab6f2a" exitCode=0 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.888940 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w4bk" event={"ID":"225e0c5d-e485-4157-88f6-bfb73efdac10","Type":"ContainerDied","Data":"815a2f97be46dbde7392b009712b24479ab068c7ef6631be841cd020cbab6f2a"} Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.892706 4719 generic.go:334] "Generic (PLEG): container finished" podID="44695712-4b17-480e-8c22-a392046822ca" containerID="1039bd2e6a283013c1914d81074e78a9b5ef9c36ee2daf4768271f309e87e1d3" exitCode=0 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.892776 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxvcb" event={"ID":"44695712-4b17-480e-8c22-a392046822ca","Type":"ContainerDied","Data":"1039bd2e6a283013c1914d81074e78a9b5ef9c36ee2daf4768271f309e87e1d3"} Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.895770 4719 generic.go:334] "Generic (PLEG): container finished" podID="5de2277b-38a8-4be5-b4f5-b7516197192c" containerID="3a6d10cfe2ade7410f3ac7ea893a64ff7cb491f5ec4e015f333f7eedd228f538" exitCode=0 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.895894 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" event={"ID":"5de2277b-38a8-4be5-b4f5-b7516197192c","Type":"ContainerDied","Data":"3a6d10cfe2ade7410f3ac7ea893a64ff7cb491f5ec4e015f333f7eedd228f538"} Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.901313 4719 generic.go:334] "Generic (PLEG): container finished" podID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerID="e55f0213163fa2e4247ff82d1df64f2e99e5ecb1286b7d3e9285074c7fa62ca3" exitCode=0 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.901393 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl7hj" event={"ID":"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc","Type":"ContainerDied","Data":"e55f0213163fa2e4247ff82d1df64f2e99e5ecb1286b7d3e9285074c7fa62ca3"} Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.905841 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" event={"ID":"e804e02c-d21b-4881-9bc8-2de04a0928f9","Type":"ContainerStarted","Data":"1ef928bc1d41a830ececb59b97767d7da4de79893999255b84ccbbdbc3ea7809"} Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.908152 4719 generic.go:334] "Generic (PLEG): container finished" podID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerID="7158d7d1dc94062b94b1ca9b2dcf965acf5e78932a2793d89d8de35a21c2d858" exitCode=0 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.908372 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nxt4c" podUID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerName="registry-server" containerID="cri-o://9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60" gracePeriod=2 Dec 15 12:30:23 crc kubenswrapper[4719]: I1215 12:30:23.908666 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fvzt6" event={"ID":"97f03079-f67f-4569-a9e3-6885d9c9e8cc","Type":"ContainerDied","Data":"7158d7d1dc94062b94b1ca9b2dcf965acf5e78932a2793d89d8de35a21c2d858"} Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.098051 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.190911 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h57pt\" (UniqueName: \"kubernetes.io/projected/97f03079-f67f-4569-a9e3-6885d9c9e8cc-kube-api-access-h57pt\") pod \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.190975 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-catalog-content\") pod \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.191041 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-utilities\") pod \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\" (UID: \"97f03079-f67f-4569-a9e3-6885d9c9e8cc\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.191815 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-utilities" (OuterVolumeSpecName: "utilities") pod "97f03079-f67f-4569-a9e3-6885d9c9e8cc" (UID: "97f03079-f67f-4569-a9e3-6885d9c9e8cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.197018 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97f03079-f67f-4569-a9e3-6885d9c9e8cc-kube-api-access-h57pt" (OuterVolumeSpecName: "kube-api-access-h57pt") pod "97f03079-f67f-4569-a9e3-6885d9c9e8cc" (UID: "97f03079-f67f-4569-a9e3-6885d9c9e8cc"). InnerVolumeSpecName "kube-api-access-h57pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.235781 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.251754 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97f03079-f67f-4569-a9e3-6885d9c9e8cc" (UID: "97f03079-f67f-4569-a9e3-6885d9c9e8cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.257386 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.257494 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.271995 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.292801 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.292840 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h57pt\" (UniqueName: \"kubernetes.io/projected/97f03079-f67f-4569-a9e3-6885d9c9e8cc-kube-api-access-h57pt\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.292863 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f03079-f67f-4569-a9e3-6885d9c9e8cc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.313222 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397032 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-utilities\") pod \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397082 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-catalog-content\") pod \"44695712-4b17-480e-8c22-a392046822ca\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397111 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db5mj\" (UniqueName: \"kubernetes.io/projected/73cbed55-0dac-4738-b4e2-0d863c20570f-kube-api-access-db5mj\") pod \"73cbed55-0dac-4738-b4e2-0d863c20570f\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397145 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-catalog-content\") pod \"73cbed55-0dac-4738-b4e2-0d863c20570f\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397174 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-catalog-content\") pod \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397201 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-catalog-content\") pod \"225e0c5d-e485-4157-88f6-bfb73efdac10\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397250 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-utilities\") pod \"73cbed55-0dac-4738-b4e2-0d863c20570f\" (UID: \"73cbed55-0dac-4738-b4e2-0d863c20570f\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397285 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-operator-metrics\") pod \"5de2277b-38a8-4be5-b4f5-b7516197192c\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397310 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-trusted-ca\") pod \"5de2277b-38a8-4be5-b4f5-b7516197192c\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397332 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8kvk\" (UniqueName: \"kubernetes.io/projected/225e0c5d-e485-4157-88f6-bfb73efdac10-kube-api-access-j8kvk\") pod \"225e0c5d-e485-4157-88f6-bfb73efdac10\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397352 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqj6v\" (UniqueName: \"kubernetes.io/projected/44695712-4b17-480e-8c22-a392046822ca-kube-api-access-jqj6v\") pod \"44695712-4b17-480e-8c22-a392046822ca\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397503 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-utilities\") pod \"225e0c5d-e485-4157-88f6-bfb73efdac10\" (UID: \"225e0c5d-e485-4157-88f6-bfb73efdac10\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397531 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-utilities\") pod \"44695712-4b17-480e-8c22-a392046822ca\" (UID: \"44695712-4b17-480e-8c22-a392046822ca\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397544 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brnw5\" (UniqueName: \"kubernetes.io/projected/5de2277b-38a8-4be5-b4f5-b7516197192c-kube-api-access-brnw5\") pod \"5de2277b-38a8-4be5-b4f5-b7516197192c\" (UID: \"5de2277b-38a8-4be5-b4f5-b7516197192c\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.397584 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-584kg\" (UniqueName: \"kubernetes.io/projected/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-kube-api-access-584kg\") pod \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\" (UID: \"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc\") " Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.398500 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-utilities" (OuterVolumeSpecName: "utilities") pod "06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" (UID: "06c7e7eb-5780-4b52-bd86-f5b74a20f8bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.399126 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "5de2277b-38a8-4be5-b4f5-b7516197192c" (UID: "5de2277b-38a8-4be5-b4f5-b7516197192c"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.405040 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44695712-4b17-480e-8c22-a392046822ca-kube-api-access-jqj6v" (OuterVolumeSpecName: "kube-api-access-jqj6v") pod "44695712-4b17-480e-8c22-a392046822ca" (UID: "44695712-4b17-480e-8c22-a392046822ca"). InnerVolumeSpecName "kube-api-access-jqj6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.407447 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-utilities" (OuterVolumeSpecName: "utilities") pod "44695712-4b17-480e-8c22-a392046822ca" (UID: "44695712-4b17-480e-8c22-a392046822ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.408664 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-utilities" (OuterVolumeSpecName: "utilities") pod "225e0c5d-e485-4157-88f6-bfb73efdac10" (UID: "225e0c5d-e485-4157-88f6-bfb73efdac10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.409821 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n2djf"] Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.410179 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.410269 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.410370 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerName="extract-content" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.410461 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerName="extract-content" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.410550 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.410619 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.410686 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerName="extract-utilities" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.410751 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerName="extract-utilities" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.410847 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de2277b-38a8-4be5-b4f5-b7516197192c" containerName="marketplace-operator" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.410941 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de2277b-38a8-4be5-b4f5-b7516197192c" containerName="marketplace-operator" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.411023 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44695712-4b17-480e-8c22-a392046822ca" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.411088 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="44695712-4b17-480e-8c22-a392046822ca" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.411151 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.411215 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.411276 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerName="extract-utilities" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.411350 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerName="extract-utilities" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.411459 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.411547 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.411598 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44695712-4b17-480e-8c22-a392046822ca" containerName="extract-content" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.411643 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="44695712-4b17-480e-8c22-a392046822ca" containerName="extract-content" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.411694 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerName="extract-content" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.411746 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerName="extract-content" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.411797 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerName="extract-content" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.411841 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerName="extract-content" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.411906 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerName="extract-content" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.411953 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerName="extract-content" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.412002 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44695712-4b17-480e-8c22-a392046822ca" containerName="extract-utilities" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.412058 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="44695712-4b17-480e-8c22-a392046822ca" containerName="extract-utilities" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.412125 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerName="extract-utilities" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.412199 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerName="extract-utilities" Dec 15 12:30:24 crc kubenswrapper[4719]: E1215 12:30:24.412262 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerName="extract-utilities" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.412310 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerName="extract-utilities" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.412487 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de2277b-38a8-4be5-b4f5-b7516197192c" containerName="marketplace-operator" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.412567 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.412621 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="44695712-4b17-480e-8c22-a392046822ca" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.412672 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.412721 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="225e0c5d-e485-4157-88f6-bfb73efdac10" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.412770 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" containerName="registry-server" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.412824 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "5de2277b-38a8-4be5-b4f5-b7516197192c" (UID: "5de2277b-38a8-4be5-b4f5-b7516197192c"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.416045 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/225e0c5d-e485-4157-88f6-bfb73efdac10-kube-api-access-j8kvk" (OuterVolumeSpecName: "kube-api-access-j8kvk") pod "225e0c5d-e485-4157-88f6-bfb73efdac10" (UID: "225e0c5d-e485-4157-88f6-bfb73efdac10"). InnerVolumeSpecName "kube-api-access-j8kvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.416829 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de2277b-38a8-4be5-b4f5-b7516197192c-kube-api-access-brnw5" (OuterVolumeSpecName: "kube-api-access-brnw5") pod "5de2277b-38a8-4be5-b4f5-b7516197192c" (UID: "5de2277b-38a8-4be5-b4f5-b7516197192c"). InnerVolumeSpecName "kube-api-access-brnw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.418368 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-kube-api-access-584kg" (OuterVolumeSpecName: "kube-api-access-584kg") pod "06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" (UID: "06c7e7eb-5780-4b52-bd86-f5b74a20f8bc"). InnerVolumeSpecName "kube-api-access-584kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.419757 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-utilities" (OuterVolumeSpecName: "utilities") pod "73cbed55-0dac-4738-b4e2-0d863c20570f" (UID: "73cbed55-0dac-4738-b4e2-0d863c20570f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.429456 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n2djf"] Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.429573 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73cbed55-0dac-4738-b4e2-0d863c20570f-kube-api-access-db5mj" (OuterVolumeSpecName: "kube-api-access-db5mj") pod "73cbed55-0dac-4738-b4e2-0d863c20570f" (UID: "73cbed55-0dac-4738-b4e2-0d863c20570f"). InnerVolumeSpecName "kube-api-access-db5mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.429683 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.486417 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44695712-4b17-480e-8c22-a392046822ca" (UID: "44695712-4b17-480e-8c22-a392046822ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498456 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-utilities\") pod \"certified-operators-n2djf\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498506 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-catalog-content\") pod \"certified-operators-n2djf\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498576 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sl94\" (UniqueName: \"kubernetes.io/projected/62b57fef-4888-4a52-9e33-c3c22873b5fb-kube-api-access-8sl94\") pod \"certified-operators-n2djf\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498642 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498657 4719 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498672 4719 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5de2277b-38a8-4be5-b4f5-b7516197192c-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498682 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8kvk\" (UniqueName: \"kubernetes.io/projected/225e0c5d-e485-4157-88f6-bfb73efdac10-kube-api-access-j8kvk\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498693 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqj6v\" (UniqueName: \"kubernetes.io/projected/44695712-4b17-480e-8c22-a392046822ca-kube-api-access-jqj6v\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498703 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498714 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498724 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brnw5\" (UniqueName: \"kubernetes.io/projected/5de2277b-38a8-4be5-b4f5-b7516197192c-kube-api-access-brnw5\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498734 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-584kg\" (UniqueName: \"kubernetes.io/projected/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-kube-api-access-584kg\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498744 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498754 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44695712-4b17-480e-8c22-a392046822ca-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.498765 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db5mj\" (UniqueName: \"kubernetes.io/projected/73cbed55-0dac-4738-b4e2-0d863c20570f-kube-api-access-db5mj\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.506174 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73cbed55-0dac-4738-b4e2-0d863c20570f" (UID: "73cbed55-0dac-4738-b4e2-0d863c20570f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.520775 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" (UID: "06c7e7eb-5780-4b52-bd86-f5b74a20f8bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.565552 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "225e0c5d-e485-4157-88f6-bfb73efdac10" (UID: "225e0c5d-e485-4157-88f6-bfb73efdac10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.599569 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-utilities\") pod \"certified-operators-n2djf\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.599625 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-catalog-content\") pod \"certified-operators-n2djf\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.599694 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sl94\" (UniqueName: \"kubernetes.io/projected/62b57fef-4888-4a52-9e33-c3c22873b5fb-kube-api-access-8sl94\") pod \"certified-operators-n2djf\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.599757 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73cbed55-0dac-4738-b4e2-0d863c20570f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.599771 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.599782 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/225e0c5d-e485-4157-88f6-bfb73efdac10-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.600310 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-utilities\") pod \"certified-operators-n2djf\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.600373 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-catalog-content\") pod \"certified-operators-n2djf\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.616396 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sl94\" (UniqueName: \"kubernetes.io/projected/62b57fef-4888-4a52-9e33-c3c22873b5fb-kube-api-access-8sl94\") pod \"certified-operators-n2djf\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.765377 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.922981 4719 generic.go:334] "Generic (PLEG): container finished" podID="73cbed55-0dac-4738-b4e2-0d863c20570f" containerID="9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60" exitCode=0 Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.923251 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxt4c" event={"ID":"73cbed55-0dac-4738-b4e2-0d863c20570f","Type":"ContainerDied","Data":"9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60"} Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.923050 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxt4c" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.923292 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxt4c" event={"ID":"73cbed55-0dac-4738-b4e2-0d863c20570f","Type":"ContainerDied","Data":"edffa6e426dff1abef5bb791bf2e4bc4019c2a6e1df2b09bf3d4d962599535ee"} Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.923315 4719 scope.go:117] "RemoveContainer" containerID="9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.929584 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fvzt6" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.929583 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fvzt6" event={"ID":"97f03079-f67f-4569-a9e3-6885d9c9e8cc","Type":"ContainerDied","Data":"73616861ef3f4044fd85912a0850e490c31c0a87b9393c75a11f1ee73d48eb1b"} Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.934227 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n2djf"] Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.934841 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w4bk" event={"ID":"225e0c5d-e485-4157-88f6-bfb73efdac10","Type":"ContainerDied","Data":"932da17b12345dcda3761fe515ade0307af0a6896827868d1bb122648faa3704"} Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.934979 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6w4bk" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.944108 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxvcb" event={"ID":"44695712-4b17-480e-8c22-a392046822ca","Type":"ContainerDied","Data":"04d88408b05bf6fc2eece9f29bd6d2fbb544d5371732aa1a7f85276682149e17"} Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.944474 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxvcb" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.947790 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" event={"ID":"5de2277b-38a8-4be5-b4f5-b7516197192c","Type":"ContainerDied","Data":"c38b86ca2e0f66f1b37faeb4e328170f3d894a950a26f1142130b29dcf3d3e09"} Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.947916 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ds7r2" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.948343 4719 scope.go:117] "RemoveContainer" containerID="6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.962922 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nxt4c"] Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.965177 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl7hj" event={"ID":"06c7e7eb-5780-4b52-bd86-f5b74a20f8bc","Type":"ContainerDied","Data":"b12997847ca460285bc1b20b78dcc282176b2b151c93435e5230fa3978e605bb"} Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.965320 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl7hj" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.969168 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nxt4c"] Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.972044 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" event={"ID":"e804e02c-d21b-4881-9bc8-2de04a0928f9","Type":"ContainerStarted","Data":"cc9f8a435ac2e6d9233fdea3aea3cefc41b8e3ce0e603de5a9c9bfcf5df3fc7b"} Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.972829 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.978903 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fvzt6"] Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.981139 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.986515 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fvzt6"] Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.996229 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6w4bk"] Dec 15 12:30:24 crc kubenswrapper[4719]: I1215 12:30:24.997459 4719 scope.go:117] "RemoveContainer" containerID="67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.000110 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6w4bk"] Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.007520 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kxrn8" podStartSLOduration=2.007481455 podStartE2EDuration="2.007481455s" podCreationTimestamp="2025-12-15 12:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:30:25.006617139 +0000 UTC m=+785.948910179" watchObservedRunningTime="2025-12-15 12:30:25.007481455 +0000 UTC m=+785.949774485" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.024728 4719 scope.go:117] "RemoveContainer" containerID="9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60" Dec 15 12:30:25 crc kubenswrapper[4719]: E1215 12:30:25.025371 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60\": container with ID starting with 9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60 not found: ID does not exist" containerID="9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.025469 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60"} err="failed to get container status \"9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60\": rpc error: code = NotFound desc = could not find container \"9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60\": container with ID starting with 9cd0199c0d0843296fd84f807bf503b9fd301d8348a4559a5109bde1a416dc60 not found: ID does not exist" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.025541 4719 scope.go:117] "RemoveContainer" containerID="6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62" Dec 15 12:30:25 crc kubenswrapper[4719]: E1215 12:30:25.025818 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62\": container with ID starting with 6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62 not found: ID does not exist" containerID="6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.026011 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62"} err="failed to get container status \"6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62\": rpc error: code = NotFound desc = could not find container \"6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62\": container with ID starting with 6d2fbd4a55261981e983ffbaa8a0992053e8073e4a90a3b663886ff43350da62 not found: ID does not exist" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.026144 4719 scope.go:117] "RemoveContainer" containerID="67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6" Dec 15 12:30:25 crc kubenswrapper[4719]: E1215 12:30:25.027250 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6\": container with ID starting with 67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6 not found: ID does not exist" containerID="67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.027374 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6"} err="failed to get container status \"67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6\": rpc error: code = NotFound desc = could not find container \"67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6\": container with ID starting with 67f0333c8d8d50505aa7bded0093dbf29db600b5c8f0043d0139b6cf88fc70e6 not found: ID does not exist" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.029952 4719 scope.go:117] "RemoveContainer" containerID="7158d7d1dc94062b94b1ca9b2dcf965acf5e78932a2793d89d8de35a21c2d858" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.051183 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xl7hj"] Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.056479 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xl7hj"] Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.059787 4719 scope.go:117] "RemoveContainer" containerID="75ec5bf5bde70701211e415aeca48c9791f96584eebf3e0bb037276f0b8044c6" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.075190 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ds7r2"] Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.081101 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ds7r2"] Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.089389 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxvcb"] Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.092202 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxvcb"] Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.092416 4719 scope.go:117] "RemoveContainer" containerID="8df605053a4ecfb71360d56b1d98a8a5b51cfbba2ff5203b5b77356528ba48d3" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.105359 4719 scope.go:117] "RemoveContainer" containerID="815a2f97be46dbde7392b009712b24479ab068c7ef6631be841cd020cbab6f2a" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.117435 4719 scope.go:117] "RemoveContainer" containerID="24b1f278b37b4e508f9f99ffcdaf35441ab94692e7b8d8da71098a51661d785d" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.134720 4719 scope.go:117] "RemoveContainer" containerID="985a03d56e1240db3ee814fcb6cf6beccc4ddc2e0c7f3c46b11fcc21b5115e9c" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.151785 4719 scope.go:117] "RemoveContainer" containerID="1039bd2e6a283013c1914d81074e78a9b5ef9c36ee2daf4768271f309e87e1d3" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.218626 4719 scope.go:117] "RemoveContainer" containerID="f39de7126b137ef785f7e2f74fa5de47afd5b9a621e4f8f2edd543f4a3d1637c" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.235378 4719 scope.go:117] "RemoveContainer" containerID="5701cbadfdb10fcc1aefb5fbff79e44a36bbdff4f3bb4916e9dab0c9c8bf96ca" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.263585 4719 scope.go:117] "RemoveContainer" containerID="3a6d10cfe2ade7410f3ac7ea893a64ff7cb491f5ec4e015f333f7eedd228f538" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.283706 4719 scope.go:117] "RemoveContainer" containerID="e55f0213163fa2e4247ff82d1df64f2e99e5ecb1286b7d3e9285074c7fa62ca3" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.306874 4719 scope.go:117] "RemoveContainer" containerID="19e1115aba9809b424422c33c912b56bdc15ed1d73113e0a7cfbd6290296332c" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.325506 4719 scope.go:117] "RemoveContainer" containerID="f9e8561e08ec1866df76a77520de09f0d2cb296c59e50127f48fe2ad0217d4fb" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.405900 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sgc7g"] Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.407365 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.410687 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.419769 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sgc7g"] Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.485435 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06c7e7eb-5780-4b52-bd86-f5b74a20f8bc" path="/var/lib/kubelet/pods/06c7e7eb-5780-4b52-bd86-f5b74a20f8bc/volumes" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.486563 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="225e0c5d-e485-4157-88f6-bfb73efdac10" path="/var/lib/kubelet/pods/225e0c5d-e485-4157-88f6-bfb73efdac10/volumes" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.487188 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44695712-4b17-480e-8c22-a392046822ca" path="/var/lib/kubelet/pods/44695712-4b17-480e-8c22-a392046822ca/volumes" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.489929 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5de2277b-38a8-4be5-b4f5-b7516197192c" path="/var/lib/kubelet/pods/5de2277b-38a8-4be5-b4f5-b7516197192c/volumes" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.490821 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73cbed55-0dac-4738-b4e2-0d863c20570f" path="/var/lib/kubelet/pods/73cbed55-0dac-4738-b4e2-0d863c20570f/volumes" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.492236 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97f03079-f67f-4569-a9e3-6885d9c9e8cc" path="/var/lib/kubelet/pods/97f03079-f67f-4569-a9e3-6885d9c9e8cc/volumes" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.509573 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9caada-f1dc-4aa5-b328-1e1442300524-utilities\") pod \"community-operators-sgc7g\" (UID: \"9d9caada-f1dc-4aa5-b328-1e1442300524\") " pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.509627 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzj59\" (UniqueName: \"kubernetes.io/projected/9d9caada-f1dc-4aa5-b328-1e1442300524-kube-api-access-mzj59\") pod \"community-operators-sgc7g\" (UID: \"9d9caada-f1dc-4aa5-b328-1e1442300524\") " pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.509654 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9caada-f1dc-4aa5-b328-1e1442300524-catalog-content\") pod \"community-operators-sgc7g\" (UID: \"9d9caada-f1dc-4aa5-b328-1e1442300524\") " pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.611128 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9caada-f1dc-4aa5-b328-1e1442300524-utilities\") pod \"community-operators-sgc7g\" (UID: \"9d9caada-f1dc-4aa5-b328-1e1442300524\") " pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.611192 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzj59\" (UniqueName: \"kubernetes.io/projected/9d9caada-f1dc-4aa5-b328-1e1442300524-kube-api-access-mzj59\") pod \"community-operators-sgc7g\" (UID: \"9d9caada-f1dc-4aa5-b328-1e1442300524\") " pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.611217 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9caada-f1dc-4aa5-b328-1e1442300524-catalog-content\") pod \"community-operators-sgc7g\" (UID: \"9d9caada-f1dc-4aa5-b328-1e1442300524\") " pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.611985 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9caada-f1dc-4aa5-b328-1e1442300524-catalog-content\") pod \"community-operators-sgc7g\" (UID: \"9d9caada-f1dc-4aa5-b328-1e1442300524\") " pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.612015 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9caada-f1dc-4aa5-b328-1e1442300524-utilities\") pod \"community-operators-sgc7g\" (UID: \"9d9caada-f1dc-4aa5-b328-1e1442300524\") " pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.629831 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzj59\" (UniqueName: \"kubernetes.io/projected/9d9caada-f1dc-4aa5-b328-1e1442300524-kube-api-access-mzj59\") pod \"community-operators-sgc7g\" (UID: \"9d9caada-f1dc-4aa5-b328-1e1442300524\") " pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.728005 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.899181 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sgc7g"] Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.985078 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgc7g" event={"ID":"9d9caada-f1dc-4aa5-b328-1e1442300524","Type":"ContainerStarted","Data":"e82c888536204bcb819a977cdd49b531b80ade2873df298f84301028b47a9cfe"} Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.989467 4719 generic.go:334] "Generic (PLEG): container finished" podID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerID="9c43f8ffbe420eb96ecfbc2b9665758f7fda4b0684ea25102a7b68c868254c59" exitCode=0 Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.989585 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2djf" event={"ID":"62b57fef-4888-4a52-9e33-c3c22873b5fb","Type":"ContainerDied","Data":"9c43f8ffbe420eb96ecfbc2b9665758f7fda4b0684ea25102a7b68c868254c59"} Dec 15 12:30:25 crc kubenswrapper[4719]: I1215 12:30:25.989611 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2djf" event={"ID":"62b57fef-4888-4a52-9e33-c3c22873b5fb","Type":"ContainerStarted","Data":"e1b4227636dc6faf3de4317e6848ad806dc9c113519ec52e16620ffb51233a1f"} Dec 15 12:30:26 crc kubenswrapper[4719]: I1215 12:30:26.808496 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nrldb"] Dec 15 12:30:26 crc kubenswrapper[4719]: I1215 12:30:26.809986 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:26 crc kubenswrapper[4719]: I1215 12:30:26.816419 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 15 12:30:26 crc kubenswrapper[4719]: I1215 12:30:26.820471 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrldb"] Dec 15 12:30:26 crc kubenswrapper[4719]: I1215 12:30:26.928439 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b-catalog-content\") pod \"redhat-marketplace-nrldb\" (UID: \"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b\") " pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:26 crc kubenswrapper[4719]: I1215 12:30:26.928746 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b-utilities\") pod \"redhat-marketplace-nrldb\" (UID: \"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b\") " pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:26 crc kubenswrapper[4719]: I1215 12:30:26.928847 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcmqn\" (UniqueName: \"kubernetes.io/projected/577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b-kube-api-access-wcmqn\") pod \"redhat-marketplace-nrldb\" (UID: \"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b\") " pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.000226 4719 generic.go:334] "Generic (PLEG): container finished" podID="9d9caada-f1dc-4aa5-b328-1e1442300524" containerID="580571b58ec15afaf91ec75d685703d5053c7ca7e6d3ea54ab548b941cd70453" exitCode=0 Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.000410 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgc7g" event={"ID":"9d9caada-f1dc-4aa5-b328-1e1442300524","Type":"ContainerDied","Data":"580571b58ec15afaf91ec75d685703d5053c7ca7e6d3ea54ab548b941cd70453"} Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.004805 4719 generic.go:334] "Generic (PLEG): container finished" podID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerID="3bfd11ed53b02fb77b1e697f587bddb9bd8b9d074d380ecb5b7ce08b3e8bb36e" exitCode=0 Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.004894 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2djf" event={"ID":"62b57fef-4888-4a52-9e33-c3c22873b5fb","Type":"ContainerDied","Data":"3bfd11ed53b02fb77b1e697f587bddb9bd8b9d074d380ecb5b7ce08b3e8bb36e"} Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.030046 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcmqn\" (UniqueName: \"kubernetes.io/projected/577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b-kube-api-access-wcmqn\") pod \"redhat-marketplace-nrldb\" (UID: \"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b\") " pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.030392 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b-catalog-content\") pod \"redhat-marketplace-nrldb\" (UID: \"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b\") " pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.030431 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b-utilities\") pod \"redhat-marketplace-nrldb\" (UID: \"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b\") " pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.030796 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b-utilities\") pod \"redhat-marketplace-nrldb\" (UID: \"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b\") " pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.031063 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b-catalog-content\") pod \"redhat-marketplace-nrldb\" (UID: \"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b\") " pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.052780 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcmqn\" (UniqueName: \"kubernetes.io/projected/577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b-kube-api-access-wcmqn\") pod \"redhat-marketplace-nrldb\" (UID: \"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b\") " pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.131681 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.209193 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jg7pl"] Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.210766 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.230058 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg7pl"] Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.322806 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nrldb"] Dec 15 12:30:27 crc kubenswrapper[4719]: W1215 12:30:27.324642 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod577da6cd_6fcd_424a_8d6c_2aa4fe6e4c2b.slice/crio-55969d7e034df80c91b4a689bdf2508aa5466e3adcea8dbe5356e13577464469 WatchSource:0}: Error finding container 55969d7e034df80c91b4a689bdf2508aa5466e3adcea8dbe5356e13577464469: Status 404 returned error can't find the container with id 55969d7e034df80c91b4a689bdf2508aa5466e3adcea8dbe5356e13577464469 Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.333743 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-catalog-content\") pod \"redhat-marketplace-jg7pl\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.333801 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-utilities\") pod \"redhat-marketplace-jg7pl\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.333889 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf9mv\" (UniqueName: \"kubernetes.io/projected/19682024-7dba-4dd7-91ac-c353eb56eacd-kube-api-access-bf9mv\") pod \"redhat-marketplace-jg7pl\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.435578 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-catalog-content\") pod \"redhat-marketplace-jg7pl\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.435615 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-utilities\") pod \"redhat-marketplace-jg7pl\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.435636 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf9mv\" (UniqueName: \"kubernetes.io/projected/19682024-7dba-4dd7-91ac-c353eb56eacd-kube-api-access-bf9mv\") pod \"redhat-marketplace-jg7pl\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.436438 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-utilities\") pod \"redhat-marketplace-jg7pl\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.437484 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-catalog-content\") pod \"redhat-marketplace-jg7pl\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.460677 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf9mv\" (UniqueName: \"kubernetes.io/projected/19682024-7dba-4dd7-91ac-c353eb56eacd-kube-api-access-bf9mv\") pod \"redhat-marketplace-jg7pl\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.531161 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.798533 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6qhp2"] Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.800089 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.807481 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.818158 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6qhp2"] Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.942213 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-utilities\") pod \"redhat-operators-6qhp2\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.942354 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqsnr\" (UniqueName: \"kubernetes.io/projected/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-kube-api-access-qqsnr\") pod \"redhat-operators-6qhp2\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.942384 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-catalog-content\") pod \"redhat-operators-6qhp2\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:27 crc kubenswrapper[4719]: I1215 12:30:27.979955 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg7pl"] Dec 15 12:30:27 crc kubenswrapper[4719]: W1215 12:30:27.981184 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19682024_7dba_4dd7_91ac_c353eb56eacd.slice/crio-f7dddcbdbb4ff6941736e4df6cfa5dcfac84a629a6198b4e6c6f89cb997e0376 WatchSource:0}: Error finding container f7dddcbdbb4ff6941736e4df6cfa5dcfac84a629a6198b4e6c6f89cb997e0376: Status 404 returned error can't find the container with id f7dddcbdbb4ff6941736e4df6cfa5dcfac84a629a6198b4e6c6f89cb997e0376 Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.016941 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg7pl" event={"ID":"19682024-7dba-4dd7-91ac-c353eb56eacd","Type":"ContainerStarted","Data":"f7dddcbdbb4ff6941736e4df6cfa5dcfac84a629a6198b4e6c6f89cb997e0376"} Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.018135 4719 generic.go:334] "Generic (PLEG): container finished" podID="577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b" containerID="9ac07737ba204f9cc078e89c6daec02356c36ac2335c7210463342a537e3c1fd" exitCode=0 Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.018178 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrldb" event={"ID":"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b","Type":"ContainerDied","Data":"9ac07737ba204f9cc078e89c6daec02356c36ac2335c7210463342a537e3c1fd"} Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.018270 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrldb" event={"ID":"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b","Type":"ContainerStarted","Data":"55969d7e034df80c91b4a689bdf2508aa5466e3adcea8dbe5356e13577464469"} Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.043183 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqsnr\" (UniqueName: \"kubernetes.io/projected/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-kube-api-access-qqsnr\") pod \"redhat-operators-6qhp2\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.043219 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-catalog-content\") pod \"redhat-operators-6qhp2\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.043242 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-utilities\") pod \"redhat-operators-6qhp2\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.043610 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-utilities\") pod \"redhat-operators-6qhp2\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.044059 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-catalog-content\") pod \"redhat-operators-6qhp2\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.063365 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqsnr\" (UniqueName: \"kubernetes.io/projected/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-kube-api-access-qqsnr\") pod \"redhat-operators-6qhp2\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.134521 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.224196 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-75t9h"] Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.226591 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.241475 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-75t9h"] Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.333635 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6qhp2"] Dec 15 12:30:28 crc kubenswrapper[4719]: W1215 12:30:28.338465 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56dfcdfa_9677_498d_8f13_eaf8cabc48e3.slice/crio-08d2af0a83ec4eb7d4a0254c84ecab632f260b7e768243862958fd64a90fe71a WatchSource:0}: Error finding container 08d2af0a83ec4eb7d4a0254c84ecab632f260b7e768243862958fd64a90fe71a: Status 404 returned error can't find the container with id 08d2af0a83ec4eb7d4a0254c84ecab632f260b7e768243862958fd64a90fe71a Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.348826 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-utilities\") pod \"redhat-operators-75t9h\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.348898 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-catalog-content\") pod \"redhat-operators-75t9h\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.348940 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4xcd\" (UniqueName: \"kubernetes.io/projected/722d9808-4408-46b8-ad44-263d0f8d3643-kube-api-access-t4xcd\") pod \"redhat-operators-75t9h\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.450190 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-catalog-content\") pod \"redhat-operators-75t9h\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.450642 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4xcd\" (UniqueName: \"kubernetes.io/projected/722d9808-4408-46b8-ad44-263d0f8d3643-kube-api-access-t4xcd\") pod \"redhat-operators-75t9h\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.450951 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-catalog-content\") pod \"redhat-operators-75t9h\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.451132 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-utilities\") pod \"redhat-operators-75t9h\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.451581 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-utilities\") pod \"redhat-operators-75t9h\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.470030 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4xcd\" (UniqueName: \"kubernetes.io/projected/722d9808-4408-46b8-ad44-263d0f8d3643-kube-api-access-t4xcd\") pod \"redhat-operators-75t9h\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: E1215 12:30:28.482367 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d9caada_f1dc_4aa5_b328_1e1442300524.slice/crio-5a80575ef0168a14497b316137f698082394e9f835a0acbb2c977e08c026297a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d9caada_f1dc_4aa5_b328_1e1442300524.slice/crio-conmon-5a80575ef0168a14497b316137f698082394e9f835a0acbb2c977e08c026297a.scope\": RecentStats: unable to find data in memory cache]" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.551010 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:28 crc kubenswrapper[4719]: I1215 12:30:28.793319 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-75t9h"] Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.024883 4719 generic.go:334] "Generic (PLEG): container finished" podID="9d9caada-f1dc-4aa5-b328-1e1442300524" containerID="5a80575ef0168a14497b316137f698082394e9f835a0acbb2c977e08c026297a" exitCode=0 Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.024940 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgc7g" event={"ID":"9d9caada-f1dc-4aa5-b328-1e1442300524","Type":"ContainerDied","Data":"5a80575ef0168a14497b316137f698082394e9f835a0acbb2c977e08c026297a"} Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.030029 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2djf" event={"ID":"62b57fef-4888-4a52-9e33-c3c22873b5fb","Type":"ContainerStarted","Data":"0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde"} Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.032807 4719 generic.go:334] "Generic (PLEG): container finished" podID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerID="55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619" exitCode=0 Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.032882 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qhp2" event={"ID":"56dfcdfa-9677-498d-8f13-eaf8cabc48e3","Type":"ContainerDied","Data":"55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619"} Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.033031 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qhp2" event={"ID":"56dfcdfa-9677-498d-8f13-eaf8cabc48e3","Type":"ContainerStarted","Data":"08d2af0a83ec4eb7d4a0254c84ecab632f260b7e768243862958fd64a90fe71a"} Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.038146 4719 generic.go:334] "Generic (PLEG): container finished" podID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerID="9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a" exitCode=0 Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.038221 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg7pl" event={"ID":"19682024-7dba-4dd7-91ac-c353eb56eacd","Type":"ContainerDied","Data":"9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a"} Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.041229 4719 generic.go:334] "Generic (PLEG): container finished" podID="722d9808-4408-46b8-ad44-263d0f8d3643" containerID="24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60" exitCode=0 Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.041272 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75t9h" event={"ID":"722d9808-4408-46b8-ad44-263d0f8d3643","Type":"ContainerDied","Data":"24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60"} Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.041325 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75t9h" event={"ID":"722d9808-4408-46b8-ad44-263d0f8d3643","Type":"ContainerStarted","Data":"224fdeaf5c0b6c786de40ef6abad22face882802a7cfb11e6a815dc3d2c83282"} Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.123490 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n2djf" podStartSLOduration=2.853464568 podStartE2EDuration="5.123472583s" podCreationTimestamp="2025-12-15 12:30:24 +0000 UTC" firstStartedPulling="2025-12-15 12:30:25.991318385 +0000 UTC m=+786.933611415" lastFinishedPulling="2025-12-15 12:30:28.2613264 +0000 UTC m=+789.203619430" observedRunningTime="2025-12-15 12:30:29.118689485 +0000 UTC m=+790.060982535" watchObservedRunningTime="2025-12-15 12:30:29.123472583 +0000 UTC m=+790.065765603" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.601912 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lhrf4"] Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.603187 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.619191 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lhrf4"] Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.668473 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbxlv\" (UniqueName: \"kubernetes.io/projected/ea4a4126-7481-417c-968b-f3f4e0b5ff6e-kube-api-access-pbxlv\") pod \"certified-operators-lhrf4\" (UID: \"ea4a4126-7481-417c-968b-f3f4e0b5ff6e\") " pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.668589 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea4a4126-7481-417c-968b-f3f4e0b5ff6e-utilities\") pod \"certified-operators-lhrf4\" (UID: \"ea4a4126-7481-417c-968b-f3f4e0b5ff6e\") " pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.668630 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea4a4126-7481-417c-968b-f3f4e0b5ff6e-catalog-content\") pod \"certified-operators-lhrf4\" (UID: \"ea4a4126-7481-417c-968b-f3f4e0b5ff6e\") " pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.770091 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea4a4126-7481-417c-968b-f3f4e0b5ff6e-utilities\") pod \"certified-operators-lhrf4\" (UID: \"ea4a4126-7481-417c-968b-f3f4e0b5ff6e\") " pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.770142 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea4a4126-7481-417c-968b-f3f4e0b5ff6e-catalog-content\") pod \"certified-operators-lhrf4\" (UID: \"ea4a4126-7481-417c-968b-f3f4e0b5ff6e\") " pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.770182 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbxlv\" (UniqueName: \"kubernetes.io/projected/ea4a4126-7481-417c-968b-f3f4e0b5ff6e-kube-api-access-pbxlv\") pod \"certified-operators-lhrf4\" (UID: \"ea4a4126-7481-417c-968b-f3f4e0b5ff6e\") " pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.770825 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea4a4126-7481-417c-968b-f3f4e0b5ff6e-utilities\") pod \"certified-operators-lhrf4\" (UID: \"ea4a4126-7481-417c-968b-f3f4e0b5ff6e\") " pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.770925 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea4a4126-7481-417c-968b-f3f4e0b5ff6e-catalog-content\") pod \"certified-operators-lhrf4\" (UID: \"ea4a4126-7481-417c-968b-f3f4e0b5ff6e\") " pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.791932 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbxlv\" (UniqueName: \"kubernetes.io/projected/ea4a4126-7481-417c-968b-f3f4e0b5ff6e-kube-api-access-pbxlv\") pod \"certified-operators-lhrf4\" (UID: \"ea4a4126-7481-417c-968b-f3f4e0b5ff6e\") " pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:29 crc kubenswrapper[4719]: I1215 12:30:29.985314 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:30 crc kubenswrapper[4719]: I1215 12:30:30.049420 4719 generic.go:334] "Generic (PLEG): container finished" podID="577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b" containerID="47d2c71c780641a1ac8c92edfe42698e5a53559eeaec005ef8fd2dd198cad2b7" exitCode=0 Dec 15 12:30:30 crc kubenswrapper[4719]: I1215 12:30:30.049490 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrldb" event={"ID":"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b","Type":"ContainerDied","Data":"47d2c71c780641a1ac8c92edfe42698e5a53559eeaec005ef8fd2dd198cad2b7"} Dec 15 12:30:30 crc kubenswrapper[4719]: I1215 12:30:30.054980 4719 generic.go:334] "Generic (PLEG): container finished" podID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerID="d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac" exitCode=0 Dec 15 12:30:30 crc kubenswrapper[4719]: I1215 12:30:30.055042 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg7pl" event={"ID":"19682024-7dba-4dd7-91ac-c353eb56eacd","Type":"ContainerDied","Data":"d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac"} Dec 15 12:30:30 crc kubenswrapper[4719]: I1215 12:30:30.064086 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgc7g" event={"ID":"9d9caada-f1dc-4aa5-b328-1e1442300524","Type":"ContainerStarted","Data":"781a2b64895679d5191a1c31e8137e9bd5d409e64e463fc1654027c58f648116"} Dec 15 12:30:30 crc kubenswrapper[4719]: I1215 12:30:30.121014 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sgc7g" podStartSLOduration=2.640854686 podStartE2EDuration="5.120996489s" podCreationTimestamp="2025-12-15 12:30:25 +0000 UTC" firstStartedPulling="2025-12-15 12:30:27.001800533 +0000 UTC m=+787.944093563" lastFinishedPulling="2025-12-15 12:30:29.481942336 +0000 UTC m=+790.424235366" observedRunningTime="2025-12-15 12:30:30.11495108 +0000 UTC m=+791.057244150" watchObservedRunningTime="2025-12-15 12:30:30.120996489 +0000 UTC m=+791.063289529" Dec 15 12:30:30 crc kubenswrapper[4719]: I1215 12:30:30.214588 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lhrf4"] Dec 15 12:30:31 crc kubenswrapper[4719]: I1215 12:30:31.071146 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75t9h" event={"ID":"722d9808-4408-46b8-ad44-263d0f8d3643","Type":"ContainerStarted","Data":"6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e"} Dec 15 12:30:31 crc kubenswrapper[4719]: I1215 12:30:31.072452 4719 generic.go:334] "Generic (PLEG): container finished" podID="ea4a4126-7481-417c-968b-f3f4e0b5ff6e" containerID="eaff485e85b91bb1ebcbfb1e279bdbdfbe82fdabd355ef1af307c063cfdda506" exitCode=0 Dec 15 12:30:31 crc kubenswrapper[4719]: I1215 12:30:31.072507 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhrf4" event={"ID":"ea4a4126-7481-417c-968b-f3f4e0b5ff6e","Type":"ContainerDied","Data":"eaff485e85b91bb1ebcbfb1e279bdbdfbe82fdabd355ef1af307c063cfdda506"} Dec 15 12:30:31 crc kubenswrapper[4719]: I1215 12:30:31.072530 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhrf4" event={"ID":"ea4a4126-7481-417c-968b-f3f4e0b5ff6e","Type":"ContainerStarted","Data":"03a267687914d47e58184112758afa84391ef1a2cb1d7ed6eb8a14760a911188"} Dec 15 12:30:31 crc kubenswrapper[4719]: I1215 12:30:31.075093 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nrldb" event={"ID":"577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b","Type":"ContainerStarted","Data":"4b1bcd5aec05f8009251614abc45b60b4e93b04301e4bb4eb08f9338cebc361c"} Dec 15 12:30:31 crc kubenswrapper[4719]: I1215 12:30:31.078294 4719 generic.go:334] "Generic (PLEG): container finished" podID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerID="5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4" exitCode=0 Dec 15 12:30:31 crc kubenswrapper[4719]: I1215 12:30:31.078328 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qhp2" event={"ID":"56dfcdfa-9677-498d-8f13-eaf8cabc48e3","Type":"ContainerDied","Data":"5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4"} Dec 15 12:30:31 crc kubenswrapper[4719]: I1215 12:30:31.080656 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg7pl" event={"ID":"19682024-7dba-4dd7-91ac-c353eb56eacd","Type":"ContainerStarted","Data":"56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8"} Dec 15 12:30:31 crc kubenswrapper[4719]: I1215 12:30:31.113975 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jg7pl" podStartSLOduration=2.428466621 podStartE2EDuration="4.113958271s" podCreationTimestamp="2025-12-15 12:30:27 +0000 UTC" firstStartedPulling="2025-12-15 12:30:29.039528451 +0000 UTC m=+789.981821481" lastFinishedPulling="2025-12-15 12:30:30.725020101 +0000 UTC m=+791.667313131" observedRunningTime="2025-12-15 12:30:31.111203156 +0000 UTC m=+792.053496186" watchObservedRunningTime="2025-12-15 12:30:31.113958271 +0000 UTC m=+792.056251301" Dec 15 12:30:31 crc kubenswrapper[4719]: I1215 12:30:31.151967 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nrldb" podStartSLOduration=2.635637546 podStartE2EDuration="5.151945544s" podCreationTimestamp="2025-12-15 12:30:26 +0000 UTC" firstStartedPulling="2025-12-15 12:30:28.030349924 +0000 UTC m=+788.972642954" lastFinishedPulling="2025-12-15 12:30:30.546657922 +0000 UTC m=+791.488950952" observedRunningTime="2025-12-15 12:30:31.147463904 +0000 UTC m=+792.089756924" watchObservedRunningTime="2025-12-15 12:30:31.151945544 +0000 UTC m=+792.094238574" Dec 15 12:30:32 crc kubenswrapper[4719]: I1215 12:30:32.093476 4719 generic.go:334] "Generic (PLEG): container finished" podID="722d9808-4408-46b8-ad44-263d0f8d3643" containerID="6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e" exitCode=0 Dec 15 12:30:32 crc kubenswrapper[4719]: I1215 12:30:32.093690 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75t9h" event={"ID":"722d9808-4408-46b8-ad44-263d0f8d3643","Type":"ContainerDied","Data":"6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e"} Dec 15 12:30:32 crc kubenswrapper[4719]: I1215 12:30:32.097029 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qhp2" event={"ID":"56dfcdfa-9677-498d-8f13-eaf8cabc48e3","Type":"ContainerStarted","Data":"0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd"} Dec 15 12:30:32 crc kubenswrapper[4719]: I1215 12:30:32.150710 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6qhp2" podStartSLOduration=2.5081940830000002 podStartE2EDuration="5.150672066s" podCreationTimestamp="2025-12-15 12:30:27 +0000 UTC" firstStartedPulling="2025-12-15 12:30:29.036684883 +0000 UTC m=+789.978977913" lastFinishedPulling="2025-12-15 12:30:31.679162866 +0000 UTC m=+792.621455896" observedRunningTime="2025-12-15 12:30:32.143664978 +0000 UTC m=+793.085957998" watchObservedRunningTime="2025-12-15 12:30:32.150672066 +0000 UTC m=+793.092965096" Dec 15 12:30:33 crc kubenswrapper[4719]: I1215 12:30:33.104655 4719 generic.go:334] "Generic (PLEG): container finished" podID="ea4a4126-7481-417c-968b-f3f4e0b5ff6e" containerID="2ae946cb6394d4e4d9c8ae07249b18c524bf3a4ef95806a66f8355365e26df27" exitCode=0 Dec 15 12:30:33 crc kubenswrapper[4719]: I1215 12:30:33.104694 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhrf4" event={"ID":"ea4a4126-7481-417c-968b-f3f4e0b5ff6e","Type":"ContainerDied","Data":"2ae946cb6394d4e4d9c8ae07249b18c524bf3a4ef95806a66f8355365e26df27"} Dec 15 12:30:33 crc kubenswrapper[4719]: I1215 12:30:33.108381 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75t9h" event={"ID":"722d9808-4408-46b8-ad44-263d0f8d3643","Type":"ContainerStarted","Data":"3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd"} Dec 15 12:30:33 crc kubenswrapper[4719]: I1215 12:30:33.150906 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-75t9h" podStartSLOduration=1.44412027 podStartE2EDuration="5.150882525s" podCreationTimestamp="2025-12-15 12:30:28 +0000 UTC" firstStartedPulling="2025-12-15 12:30:29.043226527 +0000 UTC m=+789.985519557" lastFinishedPulling="2025-12-15 12:30:32.749988782 +0000 UTC m=+793.692281812" observedRunningTime="2025-12-15 12:30:33.149192272 +0000 UTC m=+794.091485302" watchObservedRunningTime="2025-12-15 12:30:33.150882525 +0000 UTC m=+794.093175595" Dec 15 12:30:34 crc kubenswrapper[4719]: I1215 12:30:34.766295 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:34 crc kubenswrapper[4719]: I1215 12:30:34.767287 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:34 crc kubenswrapper[4719]: I1215 12:30:34.800906 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:35 crc kubenswrapper[4719]: I1215 12:30:35.137531 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhrf4" event={"ID":"ea4a4126-7481-417c-968b-f3f4e0b5ff6e","Type":"ContainerStarted","Data":"46a1bfab63e5b2b464fa8573baa3c96b705e2f1819eae201c1cee0548db20a9f"} Dec 15 12:30:35 crc kubenswrapper[4719]: I1215 12:30:35.159103 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lhrf4" podStartSLOduration=2.737018707 podStartE2EDuration="6.159082634s" podCreationTimestamp="2025-12-15 12:30:29 +0000 UTC" firstStartedPulling="2025-12-15 12:30:31.073726751 +0000 UTC m=+792.016019771" lastFinishedPulling="2025-12-15 12:30:34.495790668 +0000 UTC m=+795.438083698" observedRunningTime="2025-12-15 12:30:35.157056311 +0000 UTC m=+796.099349341" watchObservedRunningTime="2025-12-15 12:30:35.159082634 +0000 UTC m=+796.101375664" Dec 15 12:30:35 crc kubenswrapper[4719]: I1215 12:30:35.189754 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:35 crc kubenswrapper[4719]: I1215 12:30:35.728665 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:35 crc kubenswrapper[4719]: I1215 12:30:35.728712 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:35 crc kubenswrapper[4719]: I1215 12:30:35.769191 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:36 crc kubenswrapper[4719]: I1215 12:30:36.183958 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sgc7g" Dec 15 12:30:37 crc kubenswrapper[4719]: I1215 12:30:37.132994 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:37 crc kubenswrapper[4719]: I1215 12:30:37.133070 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:37 crc kubenswrapper[4719]: I1215 12:30:37.175109 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:37 crc kubenswrapper[4719]: I1215 12:30:37.214485 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nrldb" Dec 15 12:30:37 crc kubenswrapper[4719]: I1215 12:30:37.531448 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:37 crc kubenswrapper[4719]: I1215 12:30:37.531674 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:37 crc kubenswrapper[4719]: I1215 12:30:37.570361 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:37 crc kubenswrapper[4719]: I1215 12:30:37.994004 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n2djf"] Dec 15 12:30:38 crc kubenswrapper[4719]: I1215 12:30:38.135631 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:38 crc kubenswrapper[4719]: I1215 12:30:38.135758 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:38 crc kubenswrapper[4719]: I1215 12:30:38.152822 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n2djf" podUID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerName="registry-server" containerID="cri-o://0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde" gracePeriod=2 Dec 15 12:30:38 crc kubenswrapper[4719]: I1215 12:30:38.172533 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:38 crc kubenswrapper[4719]: I1215 12:30:38.191025 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:38 crc kubenswrapper[4719]: I1215 12:30:38.551155 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:38 crc kubenswrapper[4719]: I1215 12:30:38.551426 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:38 crc kubenswrapper[4719]: I1215 12:30:38.591348 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:39 crc kubenswrapper[4719]: I1215 12:30:39.195128 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:39 crc kubenswrapper[4719]: I1215 12:30:39.196495 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 12:30:39 crc kubenswrapper[4719]: I1215 12:30:39.985616 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:39 crc kubenswrapper[4719]: I1215 12:30:39.985709 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:40 crc kubenswrapper[4719]: I1215 12:30:40.026943 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:40 crc kubenswrapper[4719]: I1215 12:30:40.204908 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lhrf4" Dec 15 12:30:40 crc kubenswrapper[4719]: I1215 12:30:40.394338 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg7pl"] Dec 15 12:30:41 crc kubenswrapper[4719]: I1215 12:30:41.166690 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jg7pl" podUID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerName="registry-server" containerID="cri-o://56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8" gracePeriod=2 Dec 15 12:30:42 crc kubenswrapper[4719]: I1215 12:30:42.992069 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-75t9h"] Dec 15 12:30:44 crc kubenswrapper[4719]: E1215 12:30:44.767123 4719 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde is running failed: container process not found" containerID="0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde" cmd=["grpc_health_probe","-addr=:50051"] Dec 15 12:30:44 crc kubenswrapper[4719]: E1215 12:30:44.768128 4719 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde is running failed: container process not found" containerID="0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde" cmd=["grpc_health_probe","-addr=:50051"] Dec 15 12:30:44 crc kubenswrapper[4719]: E1215 12:30:44.768588 4719 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde is running failed: container process not found" containerID="0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde" cmd=["grpc_health_probe","-addr=:50051"] Dec 15 12:30:44 crc kubenswrapper[4719]: E1215 12:30:44.768631 4719 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-n2djf" podUID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerName="registry-server" Dec 15 12:30:47 crc kubenswrapper[4719]: E1215 12:30:47.532799 4719 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8 is running failed: container process not found" containerID="56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8" cmd=["grpc_health_probe","-addr=:50051"] Dec 15 12:30:47 crc kubenswrapper[4719]: E1215 12:30:47.533740 4719 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8 is running failed: container process not found" containerID="56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8" cmd=["grpc_health_probe","-addr=:50051"] Dec 15 12:30:47 crc kubenswrapper[4719]: E1215 12:30:47.534393 4719 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8 is running failed: container process not found" containerID="56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8" cmd=["grpc_health_probe","-addr=:50051"] Dec 15 12:30:47 crc kubenswrapper[4719]: E1215 12:30:47.534479 4719 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-jg7pl" podUID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerName="registry-server" Dec 15 12:30:47 crc kubenswrapper[4719]: I1215 12:30:47.714177 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-n2djf_62b57fef-4888-4a52-9e33-c3c22873b5fb/registry-server/0.log" Dec 15 12:30:47 crc kubenswrapper[4719]: I1215 12:30:47.715526 4719 generic.go:334] "Generic (PLEG): container finished" podID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerID="0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde" exitCode=-1 Dec 15 12:30:47 crc kubenswrapper[4719]: I1215 12:30:47.715618 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2djf" event={"ID":"62b57fef-4888-4a52-9e33-c3c22873b5fb","Type":"ContainerDied","Data":"0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde"} Dec 15 12:30:47 crc kubenswrapper[4719]: I1215 12:30:47.715981 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-75t9h" podUID="722d9808-4408-46b8-ad44-263d0f8d3643" containerName="registry-server" containerID="cri-o://3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd" gracePeriod=2 Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.036221 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.043180 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.048114 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.211969 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-catalog-content\") pod \"722d9808-4408-46b8-ad44-263d0f8d3643\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.212339 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sl94\" (UniqueName: \"kubernetes.io/projected/62b57fef-4888-4a52-9e33-c3c22873b5fb-kube-api-access-8sl94\") pod \"62b57fef-4888-4a52-9e33-c3c22873b5fb\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.212444 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf9mv\" (UniqueName: \"kubernetes.io/projected/19682024-7dba-4dd7-91ac-c353eb56eacd-kube-api-access-bf9mv\") pod \"19682024-7dba-4dd7-91ac-c353eb56eacd\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.212468 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-utilities\") pod \"62b57fef-4888-4a52-9e33-c3c22873b5fb\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.212494 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4xcd\" (UniqueName: \"kubernetes.io/projected/722d9808-4408-46b8-ad44-263d0f8d3643-kube-api-access-t4xcd\") pod \"722d9808-4408-46b8-ad44-263d0f8d3643\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.212526 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-catalog-content\") pod \"62b57fef-4888-4a52-9e33-c3c22873b5fb\" (UID: \"62b57fef-4888-4a52-9e33-c3c22873b5fb\") " Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.212559 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-catalog-content\") pod \"19682024-7dba-4dd7-91ac-c353eb56eacd\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.212587 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-utilities\") pod \"722d9808-4408-46b8-ad44-263d0f8d3643\" (UID: \"722d9808-4408-46b8-ad44-263d0f8d3643\") " Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.212618 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-utilities\") pod \"19682024-7dba-4dd7-91ac-c353eb56eacd\" (UID: \"19682024-7dba-4dd7-91ac-c353eb56eacd\") " Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.213797 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-utilities" (OuterVolumeSpecName: "utilities") pod "62b57fef-4888-4a52-9e33-c3c22873b5fb" (UID: "62b57fef-4888-4a52-9e33-c3c22873b5fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.214772 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-utilities" (OuterVolumeSpecName: "utilities") pod "722d9808-4408-46b8-ad44-263d0f8d3643" (UID: "722d9808-4408-46b8-ad44-263d0f8d3643"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.217644 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19682024-7dba-4dd7-91ac-c353eb56eacd-kube-api-access-bf9mv" (OuterVolumeSpecName: "kube-api-access-bf9mv") pod "19682024-7dba-4dd7-91ac-c353eb56eacd" (UID: "19682024-7dba-4dd7-91ac-c353eb56eacd"). InnerVolumeSpecName "kube-api-access-bf9mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.217816 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/722d9808-4408-46b8-ad44-263d0f8d3643-kube-api-access-t4xcd" (OuterVolumeSpecName: "kube-api-access-t4xcd") pod "722d9808-4408-46b8-ad44-263d0f8d3643" (UID: "722d9808-4408-46b8-ad44-263d0f8d3643"). InnerVolumeSpecName "kube-api-access-t4xcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.217910 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b57fef-4888-4a52-9e33-c3c22873b5fb-kube-api-access-8sl94" (OuterVolumeSpecName: "kube-api-access-8sl94") pod "62b57fef-4888-4a52-9e33-c3c22873b5fb" (UID: "62b57fef-4888-4a52-9e33-c3c22873b5fb"). InnerVolumeSpecName "kube-api-access-8sl94". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.218604 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-utilities" (OuterVolumeSpecName: "utilities") pod "19682024-7dba-4dd7-91ac-c353eb56eacd" (UID: "19682024-7dba-4dd7-91ac-c353eb56eacd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.236705 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19682024-7dba-4dd7-91ac-c353eb56eacd" (UID: "19682024-7dba-4dd7-91ac-c353eb56eacd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.264690 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62b57fef-4888-4a52-9e33-c3c22873b5fb" (UID: "62b57fef-4888-4a52-9e33-c3c22873b5fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.314273 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.314313 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sl94\" (UniqueName: \"kubernetes.io/projected/62b57fef-4888-4a52-9e33-c3c22873b5fb-kube-api-access-8sl94\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.314401 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf9mv\" (UniqueName: \"kubernetes.io/projected/19682024-7dba-4dd7-91ac-c353eb56eacd-kube-api-access-bf9mv\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.314415 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.314427 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4xcd\" (UniqueName: \"kubernetes.io/projected/722d9808-4408-46b8-ad44-263d0f8d3643-kube-api-access-t4xcd\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.314438 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62b57fef-4888-4a52-9e33-c3c22873b5fb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.314463 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19682024-7dba-4dd7-91ac-c353eb56eacd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.314474 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.355128 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "722d9808-4408-46b8-ad44-263d0f8d3643" (UID: "722d9808-4408-46b8-ad44-263d0f8d3643"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.415837 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/722d9808-4408-46b8-ad44-263d0f8d3643-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.724274 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n2djf" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.724275 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n2djf" event={"ID":"62b57fef-4888-4a52-9e33-c3c22873b5fb","Type":"ContainerDied","Data":"e1b4227636dc6faf3de4317e6848ad806dc9c113519ec52e16620ffb51233a1f"} Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.726178 4719 scope.go:117] "RemoveContainer" containerID="0c325ae3d1882d5b7f5616502c696f0be07f4999025b5f0e74677d95d6b4bbde" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.728429 4719 generic.go:334] "Generic (PLEG): container finished" podID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerID="56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8" exitCode=0 Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.728500 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg7pl" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.728511 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg7pl" event={"ID":"19682024-7dba-4dd7-91ac-c353eb56eacd","Type":"ContainerDied","Data":"56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8"} Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.728542 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg7pl" event={"ID":"19682024-7dba-4dd7-91ac-c353eb56eacd","Type":"ContainerDied","Data":"f7dddcbdbb4ff6941736e4df6cfa5dcfac84a629a6198b4e6c6f89cb997e0376"} Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.736019 4719 generic.go:334] "Generic (PLEG): container finished" podID="722d9808-4408-46b8-ad44-263d0f8d3643" containerID="3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd" exitCode=0 Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.736058 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75t9h" event={"ID":"722d9808-4408-46b8-ad44-263d0f8d3643","Type":"ContainerDied","Data":"3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd"} Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.736083 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75t9h" event={"ID":"722d9808-4408-46b8-ad44-263d0f8d3643","Type":"ContainerDied","Data":"224fdeaf5c0b6c786de40ef6abad22face882802a7cfb11e6a815dc3d2c83282"} Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.736140 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75t9h" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.765031 4719 scope.go:117] "RemoveContainer" containerID="3bfd11ed53b02fb77b1e697f587bddb9bd8b9d074d380ecb5b7ce08b3e8bb36e" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.773904 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg7pl"] Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.780104 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg7pl"] Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.790152 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n2djf"] Dec 15 12:30:48 crc kubenswrapper[4719]: E1215 12:30:48.797065 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62b57fef_4888_4a52_9e33_c3c22873b5fb.slice/crio-e1b4227636dc6faf3de4317e6848ad806dc9c113519ec52e16620ffb51233a1f\": RecentStats: unable to find data in memory cache]" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.808039 4719 scope.go:117] "RemoveContainer" containerID="9c43f8ffbe420eb96ecfbc2b9665758f7fda4b0684ea25102a7b68c868254c59" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.810377 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n2djf"] Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.813785 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-75t9h"] Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.817225 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-75t9h"] Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.830328 4719 scope.go:117] "RemoveContainer" containerID="56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.851072 4719 scope.go:117] "RemoveContainer" containerID="d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.865462 4719 scope.go:117] "RemoveContainer" containerID="9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.908476 4719 scope.go:117] "RemoveContainer" containerID="56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8" Dec 15 12:30:48 crc kubenswrapper[4719]: E1215 12:30:48.909072 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8\": container with ID starting with 56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8 not found: ID does not exist" containerID="56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.909144 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8"} err="failed to get container status \"56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8\": rpc error: code = NotFound desc = could not find container \"56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8\": container with ID starting with 56365a20c3a36330756c5b23cd5ddeebe1e34451a0b402ab853039b8e34c5ef8 not found: ID does not exist" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.909199 4719 scope.go:117] "RemoveContainer" containerID="d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac" Dec 15 12:30:48 crc kubenswrapper[4719]: E1215 12:30:48.909598 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac\": container with ID starting with d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac not found: ID does not exist" containerID="d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.909639 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac"} err="failed to get container status \"d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac\": rpc error: code = NotFound desc = could not find container \"d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac\": container with ID starting with d6b78b4b9e3b246a6a3ceab52ee1a6268c18bdbc15408a8fab2905696fc66dac not found: ID does not exist" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.909661 4719 scope.go:117] "RemoveContainer" containerID="9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a" Dec 15 12:30:48 crc kubenswrapper[4719]: E1215 12:30:48.909997 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a\": container with ID starting with 9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a not found: ID does not exist" containerID="9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.910020 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a"} err="failed to get container status \"9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a\": rpc error: code = NotFound desc = could not find container \"9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a\": container with ID starting with 9e4948c22ce98f6eeee88817a1d0be30b7517b893c868c5ee3726fa89f68695a not found: ID does not exist" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.910035 4719 scope.go:117] "RemoveContainer" containerID="3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.925832 4719 scope.go:117] "RemoveContainer" containerID="6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.941419 4719 scope.go:117] "RemoveContainer" containerID="24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.967348 4719 scope.go:117] "RemoveContainer" containerID="3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd" Dec 15 12:30:48 crc kubenswrapper[4719]: E1215 12:30:48.967768 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd\": container with ID starting with 3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd not found: ID does not exist" containerID="3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.967797 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd"} err="failed to get container status \"3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd\": rpc error: code = NotFound desc = could not find container \"3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd\": container with ID starting with 3f1eccdeccb19c964966138810e728216cfe6d5f7030a3948abd5603bd5691bd not found: ID does not exist" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.967838 4719 scope.go:117] "RemoveContainer" containerID="6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e" Dec 15 12:30:48 crc kubenswrapper[4719]: E1215 12:30:48.968152 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e\": container with ID starting with 6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e not found: ID does not exist" containerID="6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.968175 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e"} err="failed to get container status \"6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e\": rpc error: code = NotFound desc = could not find container \"6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e\": container with ID starting with 6abf74f1f4079e40f1417e9b340d2f04f88e1ca881fa38df27b5cde3ca64c70e not found: ID does not exist" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.968189 4719 scope.go:117] "RemoveContainer" containerID="24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60" Dec 15 12:30:48 crc kubenswrapper[4719]: E1215 12:30:48.968617 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60\": container with ID starting with 24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60 not found: ID does not exist" containerID="24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60" Dec 15 12:30:48 crc kubenswrapper[4719]: I1215 12:30:48.968670 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60"} err="failed to get container status \"24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60\": rpc error: code = NotFound desc = could not find container \"24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60\": container with ID starting with 24d42262e72d7b5be7da5056d4d26755637f392ce0f98103b30609aaddea6b60 not found: ID does not exist" Dec 15 12:30:49 crc kubenswrapper[4719]: I1215 12:30:49.490326 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19682024-7dba-4dd7-91ac-c353eb56eacd" path="/var/lib/kubelet/pods/19682024-7dba-4dd7-91ac-c353eb56eacd/volumes" Dec 15 12:30:49 crc kubenswrapper[4719]: I1215 12:30:49.492978 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62b57fef-4888-4a52-9e33-c3c22873b5fb" path="/var/lib/kubelet/pods/62b57fef-4888-4a52-9e33-c3c22873b5fb/volumes" Dec 15 12:30:49 crc kubenswrapper[4719]: I1215 12:30:49.494150 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="722d9808-4408-46b8-ad44-263d0f8d3643" path="/var/lib/kubelet/pods/722d9808-4408-46b8-ad44-263d0f8d3643/volumes" Dec 15 12:30:54 crc kubenswrapper[4719]: I1215 12:30:54.704759 4719 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.250990 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m"] Dec 15 12:31:02 crc kubenswrapper[4719]: E1215 12:31:02.251731 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerName="extract-utilities" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.251746 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerName="extract-utilities" Dec 15 12:31:02 crc kubenswrapper[4719]: E1215 12:31:02.251764 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722d9808-4408-46b8-ad44-263d0f8d3643" containerName="extract-utilities" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.251773 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="722d9808-4408-46b8-ad44-263d0f8d3643" containerName="extract-utilities" Dec 15 12:31:02 crc kubenswrapper[4719]: E1215 12:31:02.251781 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerName="registry-server" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.251789 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerName="registry-server" Dec 15 12:31:02 crc kubenswrapper[4719]: E1215 12:31:02.251801 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerName="extract-utilities" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.251808 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerName="extract-utilities" Dec 15 12:31:02 crc kubenswrapper[4719]: E1215 12:31:02.251819 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerName="registry-server" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.251826 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerName="registry-server" Dec 15 12:31:02 crc kubenswrapper[4719]: E1215 12:31:02.251836 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722d9808-4408-46b8-ad44-263d0f8d3643" containerName="extract-content" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.251844 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="722d9808-4408-46b8-ad44-263d0f8d3643" containerName="extract-content" Dec 15 12:31:02 crc kubenswrapper[4719]: E1215 12:31:02.251874 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722d9808-4408-46b8-ad44-263d0f8d3643" containerName="registry-server" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.251882 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="722d9808-4408-46b8-ad44-263d0f8d3643" containerName="registry-server" Dec 15 12:31:02 crc kubenswrapper[4719]: E1215 12:31:02.251893 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerName="extract-content" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.251900 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerName="extract-content" Dec 15 12:31:02 crc kubenswrapper[4719]: E1215 12:31:02.251914 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerName="extract-content" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.251921 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerName="extract-content" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.252037 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="19682024-7dba-4dd7-91ac-c353eb56eacd" containerName="registry-server" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.252054 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b57fef-4888-4a52-9e33-c3c22873b5fb" containerName="registry-server" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.252070 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="722d9808-4408-46b8-ad44-263d0f8d3643" containerName="registry-server" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.252909 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.256334 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.270283 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m"] Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.296256 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgmpl\" (UniqueName: \"kubernetes.io/projected/6def757d-1beb-4122-af03-45d75afd2574-kube-api-access-lgmpl\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.296319 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.296522 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.397203 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.397310 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.397360 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgmpl\" (UniqueName: \"kubernetes.io/projected/6def757d-1beb-4122-af03-45d75afd2574-kube-api-access-lgmpl\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.397667 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.398120 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.424339 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgmpl\" (UniqueName: \"kubernetes.io/projected/6def757d-1beb-4122-af03-45d75afd2574-kube-api-access-lgmpl\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:02 crc kubenswrapper[4719]: I1215 12:31:02.566842 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:03 crc kubenswrapper[4719]: I1215 12:31:03.005085 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m"] Dec 15 12:31:03 crc kubenswrapper[4719]: I1215 12:31:03.819994 4719 generic.go:334] "Generic (PLEG): container finished" podID="6def757d-1beb-4122-af03-45d75afd2574" containerID="dd18fc79a8dbaaa6dbd367a9e5912466e30015ff8a7a8b288ca66228c816a86c" exitCode=0 Dec 15 12:31:03 crc kubenswrapper[4719]: I1215 12:31:03.820097 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" event={"ID":"6def757d-1beb-4122-af03-45d75afd2574","Type":"ContainerDied","Data":"dd18fc79a8dbaaa6dbd367a9e5912466e30015ff8a7a8b288ca66228c816a86c"} Dec 15 12:31:03 crc kubenswrapper[4719]: I1215 12:31:03.820289 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" event={"ID":"6def757d-1beb-4122-af03-45d75afd2574","Type":"ContainerStarted","Data":"df49043eaf8c304c463be7ac54a3c88e8dc040f60a3c13e842de1afbda09afcd"} Dec 15 12:31:06 crc kubenswrapper[4719]: I1215 12:31:06.839945 4719 generic.go:334] "Generic (PLEG): container finished" podID="6def757d-1beb-4122-af03-45d75afd2574" containerID="b23a3250c25f573992869e2e845e1822337952c0180e49cf169a3568ce7c9b06" exitCode=0 Dec 15 12:31:06 crc kubenswrapper[4719]: I1215 12:31:06.840200 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" event={"ID":"6def757d-1beb-4122-af03-45d75afd2574","Type":"ContainerDied","Data":"b23a3250c25f573992869e2e845e1822337952c0180e49cf169a3568ce7c9b06"} Dec 15 12:31:07 crc kubenswrapper[4719]: I1215 12:31:07.854698 4719 generic.go:334] "Generic (PLEG): container finished" podID="6def757d-1beb-4122-af03-45d75afd2574" containerID="c07b7bd08e64b564c236b8efa50d13ac5ffdf9ea4ecf11ef28ffc83dea46a11f" exitCode=0 Dec 15 12:31:07 crc kubenswrapper[4719]: I1215 12:31:07.854781 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" event={"ID":"6def757d-1beb-4122-af03-45d75afd2574","Type":"ContainerDied","Data":"c07b7bd08e64b564c236b8efa50d13ac5ffdf9ea4ecf11ef28ffc83dea46a11f"} Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.096531 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.279322 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-bundle\") pod \"6def757d-1beb-4122-af03-45d75afd2574\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.279461 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgmpl\" (UniqueName: \"kubernetes.io/projected/6def757d-1beb-4122-af03-45d75afd2574-kube-api-access-lgmpl\") pod \"6def757d-1beb-4122-af03-45d75afd2574\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.279504 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-util\") pod \"6def757d-1beb-4122-af03-45d75afd2574\" (UID: \"6def757d-1beb-4122-af03-45d75afd2574\") " Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.280127 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-bundle" (OuterVolumeSpecName: "bundle") pod "6def757d-1beb-4122-af03-45d75afd2574" (UID: "6def757d-1beb-4122-af03-45d75afd2574"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.284400 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6def757d-1beb-4122-af03-45d75afd2574-kube-api-access-lgmpl" (OuterVolumeSpecName: "kube-api-access-lgmpl") pod "6def757d-1beb-4122-af03-45d75afd2574" (UID: "6def757d-1beb-4122-af03-45d75afd2574"). InnerVolumeSpecName "kube-api-access-lgmpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.291071 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-util" (OuterVolumeSpecName: "util") pod "6def757d-1beb-4122-af03-45d75afd2574" (UID: "6def757d-1beb-4122-af03-45d75afd2574"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.381572 4719 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.381614 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgmpl\" (UniqueName: \"kubernetes.io/projected/6def757d-1beb-4122-af03-45d75afd2574-kube-api-access-lgmpl\") on node \"crc\" DevicePath \"\"" Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.381635 4719 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6def757d-1beb-4122-af03-45d75afd2574-util\") on node \"crc\" DevicePath \"\"" Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.868296 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" event={"ID":"6def757d-1beb-4122-af03-45d75afd2574","Type":"ContainerDied","Data":"df49043eaf8c304c463be7ac54a3c88e8dc040f60a3c13e842de1afbda09afcd"} Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.868337 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df49043eaf8c304c463be7ac54a3c88e8dc040f60a3c13e842de1afbda09afcd" Dec 15 12:31:09 crc kubenswrapper[4719]: I1215 12:31:09.868350 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.804389 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-h4ptq"] Dec 15 12:31:13 crc kubenswrapper[4719]: E1215 12:31:13.804898 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6def757d-1beb-4122-af03-45d75afd2574" containerName="pull" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.804914 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6def757d-1beb-4122-af03-45d75afd2574" containerName="pull" Dec 15 12:31:13 crc kubenswrapper[4719]: E1215 12:31:13.804934 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6def757d-1beb-4122-af03-45d75afd2574" containerName="extract" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.804941 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6def757d-1beb-4122-af03-45d75afd2574" containerName="extract" Dec 15 12:31:13 crc kubenswrapper[4719]: E1215 12:31:13.804958 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6def757d-1beb-4122-af03-45d75afd2574" containerName="util" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.804967 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6def757d-1beb-4122-af03-45d75afd2574" containerName="util" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.805089 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="6def757d-1beb-4122-af03-45d75afd2574" containerName="extract" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.805555 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-h4ptq" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.808457 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.808839 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.809730 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-jr2wg" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.824058 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-h4ptq"] Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.849715 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn92z\" (UniqueName: \"kubernetes.io/projected/91b08102-b558-4d78-ade6-0360f6e7846a-kube-api-access-gn92z\") pod \"nmstate-operator-6769fb99d-h4ptq\" (UID: \"91b08102-b558-4d78-ade6-0360f6e7846a\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-h4ptq" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.951120 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn92z\" (UniqueName: \"kubernetes.io/projected/91b08102-b558-4d78-ade6-0360f6e7846a-kube-api-access-gn92z\") pod \"nmstate-operator-6769fb99d-h4ptq\" (UID: \"91b08102-b558-4d78-ade6-0360f6e7846a\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-h4ptq" Dec 15 12:31:13 crc kubenswrapper[4719]: I1215 12:31:13.968667 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn92z\" (UniqueName: \"kubernetes.io/projected/91b08102-b558-4d78-ade6-0360f6e7846a-kube-api-access-gn92z\") pod \"nmstate-operator-6769fb99d-h4ptq\" (UID: \"91b08102-b558-4d78-ade6-0360f6e7846a\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-h4ptq" Dec 15 12:31:14 crc kubenswrapper[4719]: I1215 12:31:14.118814 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-h4ptq" Dec 15 12:31:14 crc kubenswrapper[4719]: I1215 12:31:14.338018 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-h4ptq"] Dec 15 12:31:14 crc kubenswrapper[4719]: I1215 12:31:14.892996 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-h4ptq" event={"ID":"91b08102-b558-4d78-ade6-0360f6e7846a","Type":"ContainerStarted","Data":"e87cf58e877a6009f661840082b8d4fc21127a2ebb50a8987baf25ef4df1d949"} Dec 15 12:31:16 crc kubenswrapper[4719]: I1215 12:31:16.904776 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-h4ptq" event={"ID":"91b08102-b558-4d78-ade6-0360f6e7846a","Type":"ContainerStarted","Data":"41c8c804e6851e5dc11b6c7c5dce8621914afee685ae4b0f3ef0411f6149c235"} Dec 15 12:31:16 crc kubenswrapper[4719]: I1215 12:31:16.925158 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-6769fb99d-h4ptq" podStartSLOduration=1.748066477 podStartE2EDuration="3.925138511s" podCreationTimestamp="2025-12-15 12:31:13 +0000 UTC" firstStartedPulling="2025-12-15 12:31:14.356391391 +0000 UTC m=+835.298684421" lastFinishedPulling="2025-12-15 12:31:16.533463425 +0000 UTC m=+837.475756455" observedRunningTime="2025-12-15 12:31:16.922084576 +0000 UTC m=+837.864377616" watchObservedRunningTime="2025-12-15 12:31:16.925138511 +0000 UTC m=+837.867431541" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.232225 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc"] Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.234829 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.238682 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-4gzvh" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.247701 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc"] Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.253215 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4"] Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.253800 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.262341 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.263958 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt8ch\" (UniqueName: \"kubernetes.io/projected/1fd5707c-b363-45cc-b2b3-168817ec1526-kube-api-access-dt8ch\") pod \"nmstate-metrics-7f7f7578db-vwmdc\" (UID: \"1fd5707c-b363-45cc-b2b3-168817ec1526\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.264010 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cee112f9-d46f-410b-bfc5-5d6cff881bb4-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-kqgm4\" (UID: \"cee112f9-d46f-410b-bfc5-5d6cff881bb4\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.264060 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gvrq\" (UniqueName: \"kubernetes.io/projected/cee112f9-d46f-410b-bfc5-5d6cff881bb4-kube-api-access-9gvrq\") pod \"nmstate-webhook-f8fb84555-kqgm4\" (UID: \"cee112f9-d46f-410b-bfc5-5d6cff881bb4\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.272949 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-trxk8"] Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.273767 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.298046 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4"] Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.365348 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gvrq\" (UniqueName: \"kubernetes.io/projected/cee112f9-d46f-410b-bfc5-5d6cff881bb4-kube-api-access-9gvrq\") pod \"nmstate-webhook-f8fb84555-kqgm4\" (UID: \"cee112f9-d46f-410b-bfc5-5d6cff881bb4\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.365433 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6524d6cd-421e-415d-826b-7ec166bc3915-ovs-socket\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.365463 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkh7x\" (UniqueName: \"kubernetes.io/projected/6524d6cd-421e-415d-826b-7ec166bc3915-kube-api-access-gkh7x\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.365487 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6524d6cd-421e-415d-826b-7ec166bc3915-dbus-socket\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.365516 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt8ch\" (UniqueName: \"kubernetes.io/projected/1fd5707c-b363-45cc-b2b3-168817ec1526-kube-api-access-dt8ch\") pod \"nmstate-metrics-7f7f7578db-vwmdc\" (UID: \"1fd5707c-b363-45cc-b2b3-168817ec1526\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.365532 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6524d6cd-421e-415d-826b-7ec166bc3915-nmstate-lock\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.365562 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cee112f9-d46f-410b-bfc5-5d6cff881bb4-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-kqgm4\" (UID: \"cee112f9-d46f-410b-bfc5-5d6cff881bb4\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:22 crc kubenswrapper[4719]: E1215 12:31:22.365676 4719 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 15 12:31:22 crc kubenswrapper[4719]: E1215 12:31:22.365731 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cee112f9-d46f-410b-bfc5-5d6cff881bb4-tls-key-pair podName:cee112f9-d46f-410b-bfc5-5d6cff881bb4 nodeName:}" failed. No retries permitted until 2025-12-15 12:31:22.865714399 +0000 UTC m=+843.808007429 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/cee112f9-d46f-410b-bfc5-5d6cff881bb4-tls-key-pair") pod "nmstate-webhook-f8fb84555-kqgm4" (UID: "cee112f9-d46f-410b-bfc5-5d6cff881bb4") : secret "openshift-nmstate-webhook" not found Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.395816 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt8ch\" (UniqueName: \"kubernetes.io/projected/1fd5707c-b363-45cc-b2b3-168817ec1526-kube-api-access-dt8ch\") pod \"nmstate-metrics-7f7f7578db-vwmdc\" (UID: \"1fd5707c-b363-45cc-b2b3-168817ec1526\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.396921 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gvrq\" (UniqueName: \"kubernetes.io/projected/cee112f9-d46f-410b-bfc5-5d6cff881bb4-kube-api-access-9gvrq\") pod \"nmstate-webhook-f8fb84555-kqgm4\" (UID: \"cee112f9-d46f-410b-bfc5-5d6cff881bb4\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.430232 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk"] Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.431172 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.435874 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.436313 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.441252 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-6m5vf" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.452726 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk"] Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.466263 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6524d6cd-421e-415d-826b-7ec166bc3915-ovs-socket\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.466500 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkh7x\" (UniqueName: \"kubernetes.io/projected/6524d6cd-421e-415d-826b-7ec166bc3915-kube-api-access-gkh7x\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.466578 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsq78\" (UniqueName: \"kubernetes.io/projected/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-kube-api-access-vsq78\") pod \"nmstate-console-plugin-6ff7998486-ljnvk\" (UID: \"5b49fba8-f261-4a35-9aa0-669b7e7e70ab\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.466656 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6524d6cd-421e-415d-826b-7ec166bc3915-dbus-socket\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.466728 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6524d6cd-421e-415d-826b-7ec166bc3915-nmstate-lock\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.466812 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-ljnvk\" (UID: \"5b49fba8-f261-4a35-9aa0-669b7e7e70ab\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.466903 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-ljnvk\" (UID: \"5b49fba8-f261-4a35-9aa0-669b7e7e70ab\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.466440 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6524d6cd-421e-415d-826b-7ec166bc3915-ovs-socket\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.467591 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6524d6cd-421e-415d-826b-7ec166bc3915-nmstate-lock\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.467621 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6524d6cd-421e-415d-826b-7ec166bc3915-dbus-socket\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.507429 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkh7x\" (UniqueName: \"kubernetes.io/projected/6524d6cd-421e-415d-826b-7ec166bc3915-kube-api-access-gkh7x\") pod \"nmstate-handler-trxk8\" (UID: \"6524d6cd-421e-415d-826b-7ec166bc3915\") " pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.548739 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.567766 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsq78\" (UniqueName: \"kubernetes.io/projected/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-kube-api-access-vsq78\") pod \"nmstate-console-plugin-6ff7998486-ljnvk\" (UID: \"5b49fba8-f261-4a35-9aa0-669b7e7e70ab\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.568106 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-ljnvk\" (UID: \"5b49fba8-f261-4a35-9aa0-669b7e7e70ab\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.569113 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-ljnvk\" (UID: \"5b49fba8-f261-4a35-9aa0-669b7e7e70ab\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:22 crc kubenswrapper[4719]: E1215 12:31:22.569351 4719 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 15 12:31:22 crc kubenswrapper[4719]: E1215 12:31:22.569474 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-plugin-serving-cert podName:5b49fba8-f261-4a35-9aa0-669b7e7e70ab nodeName:}" failed. No retries permitted until 2025-12-15 12:31:23.069459517 +0000 UTC m=+844.011752547 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-plugin-serving-cert") pod "nmstate-console-plugin-6ff7998486-ljnvk" (UID: "5b49fba8-f261-4a35-9aa0-669b7e7e70ab") : secret "plugin-serving-cert" not found Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.569062 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-ljnvk\" (UID: \"5b49fba8-f261-4a35-9aa0-669b7e7e70ab\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.591115 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.617987 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsq78\" (UniqueName: \"kubernetes.io/projected/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-kube-api-access-vsq78\") pod \"nmstate-console-plugin-6ff7998486-ljnvk\" (UID: \"5b49fba8-f261-4a35-9aa0-669b7e7e70ab\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:22 crc kubenswrapper[4719]: W1215 12:31:22.657391 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6524d6cd_421e_415d_826b_7ec166bc3915.slice/crio-148f231a4e41e6dbc83c519e139872dad0285d00ff95b75dafde2acd13c86c53 WatchSource:0}: Error finding container 148f231a4e41e6dbc83c519e139872dad0285d00ff95b75dafde2acd13c86c53: Status 404 returned error can't find the container with id 148f231a4e41e6dbc83c519e139872dad0285d00ff95b75dafde2acd13c86c53 Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.762589 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5cf7dbcc69-vhz75"] Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.763675 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.811023 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5cf7dbcc69-vhz75"] Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.872366 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9w4t\" (UniqueName: \"kubernetes.io/projected/0340c677-46c9-46db-bc44-20372c6a4248-kube-api-access-s9w4t\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.872406 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-console-config\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.872446 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0340c677-46c9-46db-bc44-20372c6a4248-console-serving-cert\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.872465 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0340c677-46c9-46db-bc44-20372c6a4248-console-oauth-config\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.872540 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-oauth-serving-cert\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.872643 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cee112f9-d46f-410b-bfc5-5d6cff881bb4-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-kqgm4\" (UID: \"cee112f9-d46f-410b-bfc5-5d6cff881bb4\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.872681 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-trusted-ca-bundle\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.872730 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-service-ca\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.876044 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cee112f9-d46f-410b-bfc5-5d6cff881bb4-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-kqgm4\" (UID: \"cee112f9-d46f-410b-bfc5-5d6cff881bb4\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.938541 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-trxk8" event={"ID":"6524d6cd-421e-415d-826b-7ec166bc3915","Type":"ContainerStarted","Data":"148f231a4e41e6dbc83c519e139872dad0285d00ff95b75dafde2acd13c86c53"} Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.974439 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-service-ca\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.974497 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9w4t\" (UniqueName: \"kubernetes.io/projected/0340c677-46c9-46db-bc44-20372c6a4248-kube-api-access-s9w4t\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.974530 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-console-config\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.974567 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0340c677-46c9-46db-bc44-20372c6a4248-console-serving-cert\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.974588 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0340c677-46c9-46db-bc44-20372c6a4248-console-oauth-config\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.974620 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-oauth-serving-cert\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.974654 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-trusted-ca-bundle\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.975944 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-trusted-ca-bundle\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.976181 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-oauth-serving-cert\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.976314 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-console-config\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.977335 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0340c677-46c9-46db-bc44-20372c6a4248-service-ca\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.978837 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0340c677-46c9-46db-bc44-20372c6a4248-console-oauth-config\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.981015 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0340c677-46c9-46db-bc44-20372c6a4248-console-serving-cert\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:22 crc kubenswrapper[4719]: I1215 12:31:22.997140 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9w4t\" (UniqueName: \"kubernetes.io/projected/0340c677-46c9-46db-bc44-20372c6a4248-kube-api-access-s9w4t\") pod \"console-5cf7dbcc69-vhz75\" (UID: \"0340c677-46c9-46db-bc44-20372c6a4248\") " pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.075877 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc"] Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.075919 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-ljnvk\" (UID: \"5b49fba8-f261-4a35-9aa0-669b7e7e70ab\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.079769 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5b49fba8-f261-4a35-9aa0-669b7e7e70ab-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-ljnvk\" (UID: \"5b49fba8-f261-4a35-9aa0-669b7e7e70ab\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:23 crc kubenswrapper[4719]: W1215 12:31:23.086256 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fd5707c_b363_45cc_b2b3_168817ec1526.slice/crio-abc93579bee215949b52a920c20f5901484ccb2a637ac910ea734ff88c1cb578 WatchSource:0}: Error finding container abc93579bee215949b52a920c20f5901484ccb2a637ac910ea734ff88c1cb578: Status 404 returned error can't find the container with id abc93579bee215949b52a920c20f5901484ccb2a637ac910ea734ff88c1cb578 Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.095278 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.167380 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.287435 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5cf7dbcc69-vhz75"] Dec 15 12:31:23 crc kubenswrapper[4719]: W1215 12:31:23.300935 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0340c677_46c9_46db_bc44_20372c6a4248.slice/crio-3b1407616263e87613ebf8dc79872f9f31919f81e0fc6f84f74cee431ed6b250 WatchSource:0}: Error finding container 3b1407616263e87613ebf8dc79872f9f31919f81e0fc6f84f74cee431ed6b250: Status 404 returned error can't find the container with id 3b1407616263e87613ebf8dc79872f9f31919f81e0fc6f84f74cee431ed6b250 Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.349754 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.378580 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4"] Dec 15 12:31:23 crc kubenswrapper[4719]: W1215 12:31:23.384359 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcee112f9_d46f_410b_bfc5_5d6cff881bb4.slice/crio-3498c90486ce65c6a9b6120dd67c057b7e7517b65935a82cb07afc3375903ed8 WatchSource:0}: Error finding container 3498c90486ce65c6a9b6120dd67c057b7e7517b65935a82cb07afc3375903ed8: Status 404 returned error can't find the container with id 3498c90486ce65c6a9b6120dd67c057b7e7517b65935a82cb07afc3375903ed8 Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.537148 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk"] Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.948709 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" event={"ID":"5b49fba8-f261-4a35-9aa0-669b7e7e70ab","Type":"ContainerStarted","Data":"151f45969c94a6a36b2a8eea3c3d52748ed33ae8d557b6972200964c60c49876"} Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.950665 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" event={"ID":"cee112f9-d46f-410b-bfc5-5d6cff881bb4","Type":"ContainerStarted","Data":"3498c90486ce65c6a9b6120dd67c057b7e7517b65935a82cb07afc3375903ed8"} Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.951901 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc" event={"ID":"1fd5707c-b363-45cc-b2b3-168817ec1526","Type":"ContainerStarted","Data":"abc93579bee215949b52a920c20f5901484ccb2a637ac910ea734ff88c1cb578"} Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.954031 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5cf7dbcc69-vhz75" event={"ID":"0340c677-46c9-46db-bc44-20372c6a4248","Type":"ContainerStarted","Data":"41ba32cf42ed8cbeb438be0f885ae78cc6b0cfa1ac712b5db1090ecd2b883d59"} Dec 15 12:31:23 crc kubenswrapper[4719]: I1215 12:31:23.954093 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5cf7dbcc69-vhz75" event={"ID":"0340c677-46c9-46db-bc44-20372c6a4248","Type":"ContainerStarted","Data":"3b1407616263e87613ebf8dc79872f9f31919f81e0fc6f84f74cee431ed6b250"} Dec 15 12:31:25 crc kubenswrapper[4719]: I1215 12:31:25.965296 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" event={"ID":"5b49fba8-f261-4a35-9aa0-669b7e7e70ab","Type":"ContainerStarted","Data":"80cb8273d86799b4bf30098e8338f67c42fe547494d47016b728fd487a214f37"} Dec 15 12:31:25 crc kubenswrapper[4719]: I1215 12:31:25.967532 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc" event={"ID":"1fd5707c-b363-45cc-b2b3-168817ec1526","Type":"ContainerStarted","Data":"4108ed74570c18a5ecb29eb7385104ac1e514e4d9603cdf3b948621178103afd"} Dec 15 12:31:25 crc kubenswrapper[4719]: I1215 12:31:25.969267 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-trxk8" event={"ID":"6524d6cd-421e-415d-826b-7ec166bc3915","Type":"ContainerStarted","Data":"6277c20cde23b31f2dee9d524a7b8f10a8bc5fbd22ab7a1fe631e659cfaaebe0"} Dec 15 12:31:25 crc kubenswrapper[4719]: I1215 12:31:25.969970 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:25 crc kubenswrapper[4719]: I1215 12:31:25.981053 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5cf7dbcc69-vhz75" podStartSLOduration=3.981029604 podStartE2EDuration="3.981029604s" podCreationTimestamp="2025-12-15 12:31:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:31:23.981012614 +0000 UTC m=+844.923305674" watchObservedRunningTime="2025-12-15 12:31:25.981029604 +0000 UTC m=+846.923322654" Dec 15 12:31:26 crc kubenswrapper[4719]: I1215 12:31:26.000814 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-trxk8" podStartSLOduration=1.01071146 podStartE2EDuration="4.000796852s" podCreationTimestamp="2025-12-15 12:31:22 +0000 UTC" firstStartedPulling="2025-12-15 12:31:22.667198529 +0000 UTC m=+843.609491559" lastFinishedPulling="2025-12-15 12:31:25.657283921 +0000 UTC m=+846.599576951" observedRunningTime="2025-12-15 12:31:25.999905814 +0000 UTC m=+846.942198864" watchObservedRunningTime="2025-12-15 12:31:26.000796852 +0000 UTC m=+846.943089882" Dec 15 12:31:26 crc kubenswrapper[4719]: I1215 12:31:26.004056 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-ljnvk" podStartSLOduration=1.892082061 podStartE2EDuration="4.004041752s" podCreationTimestamp="2025-12-15 12:31:22 +0000 UTC" firstStartedPulling="2025-12-15 12:31:23.545599878 +0000 UTC m=+844.487892908" lastFinishedPulling="2025-12-15 12:31:25.657559559 +0000 UTC m=+846.599852599" observedRunningTime="2025-12-15 12:31:25.985174241 +0000 UTC m=+846.927467291" watchObservedRunningTime="2025-12-15 12:31:26.004041752 +0000 UTC m=+846.946334782" Dec 15 12:31:26 crc kubenswrapper[4719]: I1215 12:31:26.977137 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" event={"ID":"cee112f9-d46f-410b-bfc5-5d6cff881bb4","Type":"ContainerStarted","Data":"08d6aee531773b3abe23607eed97fa9a04c0a48592fbed8549ea0c4168801697"} Dec 15 12:31:26 crc kubenswrapper[4719]: I1215 12:31:26.997510 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" podStartSLOduration=2.370770438 podStartE2EDuration="4.997490985s" podCreationTimestamp="2025-12-15 12:31:22 +0000 UTC" firstStartedPulling="2025-12-15 12:31:23.386277219 +0000 UTC m=+844.328570249" lastFinishedPulling="2025-12-15 12:31:26.012997776 +0000 UTC m=+846.955290796" observedRunningTime="2025-12-15 12:31:26.994765791 +0000 UTC m=+847.937058841" watchObservedRunningTime="2025-12-15 12:31:26.997490985 +0000 UTC m=+847.939784015" Dec 15 12:31:27 crc kubenswrapper[4719]: I1215 12:31:27.983079 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:28 crc kubenswrapper[4719]: I1215 12:31:28.988553 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc" event={"ID":"1fd5707c-b363-45cc-b2b3-168817ec1526","Type":"ContainerStarted","Data":"89a5aea24649e0c6045aff0949549c3b51f63597bbfea814e47e08a6ba4d845c"} Dec 15 12:31:29 crc kubenswrapper[4719]: I1215 12:31:29.006270 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-vwmdc" podStartSLOduration=1.9924875370000001 podStartE2EDuration="7.006244023s" podCreationTimestamp="2025-12-15 12:31:22 +0000 UTC" firstStartedPulling="2025-12-15 12:31:23.08867927 +0000 UTC m=+844.030972300" lastFinishedPulling="2025-12-15 12:31:28.102435756 +0000 UTC m=+849.044728786" observedRunningTime="2025-12-15 12:31:29.003775007 +0000 UTC m=+849.946068047" watchObservedRunningTime="2025-12-15 12:31:29.006244023 +0000 UTC m=+849.948537083" Dec 15 12:31:32 crc kubenswrapper[4719]: I1215 12:31:32.619055 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-trxk8" Dec 15 12:31:33 crc kubenswrapper[4719]: I1215 12:31:33.095731 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:33 crc kubenswrapper[4719]: I1215 12:31:33.095788 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:33 crc kubenswrapper[4719]: I1215 12:31:33.101781 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:34 crc kubenswrapper[4719]: I1215 12:31:34.023483 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5cf7dbcc69-vhz75" Dec 15 12:31:34 crc kubenswrapper[4719]: I1215 12:31:34.088239 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-2xxrl"] Dec 15 12:31:43 crc kubenswrapper[4719]: I1215 12:31:43.175979 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-f8fb84555-kqgm4" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.195590 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr"] Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.197535 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.199351 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.212358 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr"] Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.321904 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.321970 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.322329 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzsxg\" (UniqueName: \"kubernetes.io/projected/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-kube-api-access-lzsxg\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.424552 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.424685 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzsxg\" (UniqueName: \"kubernetes.io/projected/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-kube-api-access-lzsxg\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.424757 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.425338 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.425473 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.450125 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzsxg\" (UniqueName: \"kubernetes.io/projected/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-kube-api-access-lzsxg\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.533381 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:31:55 crc kubenswrapper[4719]: I1215 12:31:55.932693 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr"] Dec 15 12:31:56 crc kubenswrapper[4719]: I1215 12:31:56.148774 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" event={"ID":"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0","Type":"ContainerStarted","Data":"687776f93ca5eaf133cfbffbe58cabb41f8701fa99c5756dcf3f74af8a47181a"} Dec 15 12:31:57 crc kubenswrapper[4719]: I1215 12:31:57.164924 4719 generic.go:334] "Generic (PLEG): container finished" podID="2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" containerID="952f0e733a4485c4ab8eacb6084be0bbc8b994b73c410a59182be94e69378ca0" exitCode=0 Dec 15 12:31:57 crc kubenswrapper[4719]: I1215 12:31:57.165133 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" event={"ID":"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0","Type":"ContainerDied","Data":"952f0e733a4485c4ab8eacb6084be0bbc8b994b73c410a59182be94e69378ca0"} Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.126979 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-2xxrl" podUID="c8458395-1102-446b-8b75-752820a9cfdb" containerName="console" containerID="cri-o://01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13" gracePeriod=15 Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.181007 4719 generic.go:334] "Generic (PLEG): container finished" podID="2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" containerID="bc7ca02f6157e00e4d5568ef05fa033bed1103c4a944d5286b8258fff525d956" exitCode=0 Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.181040 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" event={"ID":"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0","Type":"ContainerDied","Data":"bc7ca02f6157e00e4d5568ef05fa033bed1103c4a944d5286b8258fff525d956"} Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.201490 4719 patch_prober.go:28] interesting pod/console-f9d7485db-2xxrl container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.201543 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-2xxrl" podUID="c8458395-1102-446b-8b75-752820a9cfdb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.469881 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-2xxrl_c8458395-1102-446b-8b75-752820a9cfdb/console/0.log" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.469944 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.580023 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-service-ca\") pod \"c8458395-1102-446b-8b75-752820a9cfdb\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.580105 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-oauth-serving-cert\") pod \"c8458395-1102-446b-8b75-752820a9cfdb\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.580154 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-trusted-ca-bundle\") pod \"c8458395-1102-446b-8b75-752820a9cfdb\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.580192 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-oauth-config\") pod \"c8458395-1102-446b-8b75-752820a9cfdb\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.580242 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-console-config\") pod \"c8458395-1102-446b-8b75-752820a9cfdb\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.580273 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrffv\" (UniqueName: \"kubernetes.io/projected/c8458395-1102-446b-8b75-752820a9cfdb-kube-api-access-vrffv\") pod \"c8458395-1102-446b-8b75-752820a9cfdb\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.580325 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-serving-cert\") pod \"c8458395-1102-446b-8b75-752820a9cfdb\" (UID: \"c8458395-1102-446b-8b75-752820a9cfdb\") " Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.581198 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c8458395-1102-446b-8b75-752820a9cfdb" (UID: "c8458395-1102-446b-8b75-752820a9cfdb"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.581282 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-console-config" (OuterVolumeSpecName: "console-config") pod "c8458395-1102-446b-8b75-752820a9cfdb" (UID: "c8458395-1102-446b-8b75-752820a9cfdb"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.581433 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-service-ca" (OuterVolumeSpecName: "service-ca") pod "c8458395-1102-446b-8b75-752820a9cfdb" (UID: "c8458395-1102-446b-8b75-752820a9cfdb"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.582253 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c8458395-1102-446b-8b75-752820a9cfdb" (UID: "c8458395-1102-446b-8b75-752820a9cfdb"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.586531 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c8458395-1102-446b-8b75-752820a9cfdb" (UID: "c8458395-1102-446b-8b75-752820a9cfdb"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.586704 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8458395-1102-446b-8b75-752820a9cfdb-kube-api-access-vrffv" (OuterVolumeSpecName: "kube-api-access-vrffv") pod "c8458395-1102-446b-8b75-752820a9cfdb" (UID: "c8458395-1102-446b-8b75-752820a9cfdb"). InnerVolumeSpecName "kube-api-access-vrffv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.586913 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c8458395-1102-446b-8b75-752820a9cfdb" (UID: "c8458395-1102-446b-8b75-752820a9cfdb"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.683115 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrffv\" (UniqueName: \"kubernetes.io/projected/c8458395-1102-446b-8b75-752820a9cfdb-kube-api-access-vrffv\") on node \"crc\" DevicePath \"\"" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.683151 4719 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.683162 4719 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-service-ca\") on node \"crc\" DevicePath \"\"" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.683172 4719 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.683182 4719 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.683190 4719 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c8458395-1102-446b-8b75-752820a9cfdb-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:31:59 crc kubenswrapper[4719]: I1215 12:31:59.683199 4719 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c8458395-1102-446b-8b75-752820a9cfdb-console-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.189045 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-2xxrl_c8458395-1102-446b-8b75-752820a9cfdb/console/0.log" Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.189094 4719 generic.go:334] "Generic (PLEG): container finished" podID="c8458395-1102-446b-8b75-752820a9cfdb" containerID="01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13" exitCode=2 Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.189150 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2xxrl" event={"ID":"c8458395-1102-446b-8b75-752820a9cfdb","Type":"ContainerDied","Data":"01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13"} Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.189179 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2xxrl" event={"ID":"c8458395-1102-446b-8b75-752820a9cfdb","Type":"ContainerDied","Data":"3d9bbd2d1af5dac6cdaf45afb1ed3c8f062c0588580ca88a3f1283273ec433b2"} Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.189233 4719 scope.go:117] "RemoveContainer" containerID="01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13" Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.189269 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2xxrl" Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.194261 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" event={"ID":"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0","Type":"ContainerDied","Data":"f80da15144a7ba4b3520ffb60ed82eef6464897e3912a931602129054f7cbb1e"} Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.194109 4719 generic.go:334] "Generic (PLEG): container finished" podID="2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" containerID="f80da15144a7ba4b3520ffb60ed82eef6464897e3912a931602129054f7cbb1e" exitCode=0 Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.210687 4719 scope.go:117] "RemoveContainer" containerID="01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13" Dec 15 12:32:00 crc kubenswrapper[4719]: E1215 12:32:00.213880 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13\": container with ID starting with 01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13 not found: ID does not exist" containerID="01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13" Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.213922 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13"} err="failed to get container status \"01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13\": rpc error: code = NotFound desc = could not find container \"01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13\": container with ID starting with 01902151e2352b8d577a2cecee4cdf5338c4eeecdbf0382a712ddfe984247b13 not found: ID does not exist" Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.240441 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-2xxrl"] Dec 15 12:32:00 crc kubenswrapper[4719]: I1215 12:32:00.243516 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-2xxrl"] Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.485330 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8458395-1102-446b-8b75-752820a9cfdb" path="/var/lib/kubelet/pods/c8458395-1102-446b-8b75-752820a9cfdb/volumes" Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.511973 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.608401 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzsxg\" (UniqueName: \"kubernetes.io/projected/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-kube-api-access-lzsxg\") pod \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.608461 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-bundle\") pod \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.608504 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-util\") pod \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\" (UID: \"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0\") " Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.610501 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-bundle" (OuterVolumeSpecName: "bundle") pod "2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" (UID: "2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.616057 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-kube-api-access-lzsxg" (OuterVolumeSpecName: "kube-api-access-lzsxg") pod "2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" (UID: "2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0"). InnerVolumeSpecName "kube-api-access-lzsxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.622162 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-util" (OuterVolumeSpecName: "util") pod "2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" (UID: "2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.717815 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzsxg\" (UniqueName: \"kubernetes.io/projected/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-kube-api-access-lzsxg\") on node \"crc\" DevicePath \"\"" Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.717875 4719 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:32:01 crc kubenswrapper[4719]: I1215 12:32:01.717888 4719 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0-util\") on node \"crc\" DevicePath \"\"" Dec 15 12:32:02 crc kubenswrapper[4719]: I1215 12:32:02.211197 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" event={"ID":"2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0","Type":"ContainerDied","Data":"687776f93ca5eaf133cfbffbe58cabb41f8701fa99c5756dcf3f74af8a47181a"} Dec 15 12:32:02 crc kubenswrapper[4719]: I1215 12:32:02.211248 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr" Dec 15 12:32:02 crc kubenswrapper[4719]: I1215 12:32:02.211282 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="687776f93ca5eaf133cfbffbe58cabb41f8701fa99c5756dcf3f74af8a47181a" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.104045 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck"] Dec 15 12:32:11 crc kubenswrapper[4719]: E1215 12:32:11.104988 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8458395-1102-446b-8b75-752820a9cfdb" containerName="console" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.105001 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8458395-1102-446b-8b75-752820a9cfdb" containerName="console" Dec 15 12:32:11 crc kubenswrapper[4719]: E1215 12:32:11.105011 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" containerName="pull" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.105017 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" containerName="pull" Dec 15 12:32:11 crc kubenswrapper[4719]: E1215 12:32:11.105033 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" containerName="util" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.105039 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" containerName="util" Dec 15 12:32:11 crc kubenswrapper[4719]: E1215 12:32:11.105046 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" containerName="extract" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.105051 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" containerName="extract" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.105137 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8458395-1102-446b-8b75-752820a9cfdb" containerName="console" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.105147 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0" containerName="extract" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.105505 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.107430 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.109274 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.109408 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.115982 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.120424 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-2zpfc" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.131915 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck"] Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.237919 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5009ea4d-5e13-4ac3-ad25-8c0976b0aafb-apiservice-cert\") pod \"metallb-operator-controller-manager-6cf5cc9b68-dpxck\" (UID: \"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb\") " pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.237964 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5009ea4d-5e13-4ac3-ad25-8c0976b0aafb-webhook-cert\") pod \"metallb-operator-controller-manager-6cf5cc9b68-dpxck\" (UID: \"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb\") " pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.238026 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4bpq\" (UniqueName: \"kubernetes.io/projected/5009ea4d-5e13-4ac3-ad25-8c0976b0aafb-kube-api-access-t4bpq\") pod \"metallb-operator-controller-manager-6cf5cc9b68-dpxck\" (UID: \"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb\") " pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.338715 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5009ea4d-5e13-4ac3-ad25-8c0976b0aafb-apiservice-cert\") pod \"metallb-operator-controller-manager-6cf5cc9b68-dpxck\" (UID: \"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb\") " pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.338773 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5009ea4d-5e13-4ac3-ad25-8c0976b0aafb-webhook-cert\") pod \"metallb-operator-controller-manager-6cf5cc9b68-dpxck\" (UID: \"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb\") " pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.338868 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4bpq\" (UniqueName: \"kubernetes.io/projected/5009ea4d-5e13-4ac3-ad25-8c0976b0aafb-kube-api-access-t4bpq\") pod \"metallb-operator-controller-manager-6cf5cc9b68-dpxck\" (UID: \"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb\") " pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.344619 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5009ea4d-5e13-4ac3-ad25-8c0976b0aafb-webhook-cert\") pod \"metallb-operator-controller-manager-6cf5cc9b68-dpxck\" (UID: \"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb\") " pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.345096 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5009ea4d-5e13-4ac3-ad25-8c0976b0aafb-apiservice-cert\") pod \"metallb-operator-controller-manager-6cf5cc9b68-dpxck\" (UID: \"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb\") " pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.372716 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4bpq\" (UniqueName: \"kubernetes.io/projected/5009ea4d-5e13-4ac3-ad25-8c0976b0aafb-kube-api-access-t4bpq\") pod \"metallb-operator-controller-manager-6cf5cc9b68-dpxck\" (UID: \"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb\") " pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.408722 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r"] Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.409370 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.412312 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.412313 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.412376 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-2nbhr" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.419651 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.430308 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r"] Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.542641 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/15904dfe-b1ca-4922-95cf-6f3724fae181-webhook-cert\") pod \"metallb-operator-webhook-server-76fbf99b47-d626r\" (UID: \"15904dfe-b1ca-4922-95cf-6f3724fae181\") " pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.542865 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/15904dfe-b1ca-4922-95cf-6f3724fae181-apiservice-cert\") pod \"metallb-operator-webhook-server-76fbf99b47-d626r\" (UID: \"15904dfe-b1ca-4922-95cf-6f3724fae181\") " pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.542896 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zxx9\" (UniqueName: \"kubernetes.io/projected/15904dfe-b1ca-4922-95cf-6f3724fae181-kube-api-access-4zxx9\") pod \"metallb-operator-webhook-server-76fbf99b47-d626r\" (UID: \"15904dfe-b1ca-4922-95cf-6f3724fae181\") " pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.644395 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zxx9\" (UniqueName: \"kubernetes.io/projected/15904dfe-b1ca-4922-95cf-6f3724fae181-kube-api-access-4zxx9\") pod \"metallb-operator-webhook-server-76fbf99b47-d626r\" (UID: \"15904dfe-b1ca-4922-95cf-6f3724fae181\") " pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.644494 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/15904dfe-b1ca-4922-95cf-6f3724fae181-webhook-cert\") pod \"metallb-operator-webhook-server-76fbf99b47-d626r\" (UID: \"15904dfe-b1ca-4922-95cf-6f3724fae181\") " pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.644524 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/15904dfe-b1ca-4922-95cf-6f3724fae181-apiservice-cert\") pod \"metallb-operator-webhook-server-76fbf99b47-d626r\" (UID: \"15904dfe-b1ca-4922-95cf-6f3724fae181\") " pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.651299 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/15904dfe-b1ca-4922-95cf-6f3724fae181-webhook-cert\") pod \"metallb-operator-webhook-server-76fbf99b47-d626r\" (UID: \"15904dfe-b1ca-4922-95cf-6f3724fae181\") " pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.651847 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/15904dfe-b1ca-4922-95cf-6f3724fae181-apiservice-cert\") pod \"metallb-operator-webhook-server-76fbf99b47-d626r\" (UID: \"15904dfe-b1ca-4922-95cf-6f3724fae181\") " pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.675510 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zxx9\" (UniqueName: \"kubernetes.io/projected/15904dfe-b1ca-4922-95cf-6f3724fae181-kube-api-access-4zxx9\") pod \"metallb-operator-webhook-server-76fbf99b47-d626r\" (UID: \"15904dfe-b1ca-4922-95cf-6f3724fae181\") " pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.723309 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.865035 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck"] Dec 15 12:32:11 crc kubenswrapper[4719]: I1215 12:32:11.965912 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r"] Dec 15 12:32:11 crc kubenswrapper[4719]: W1215 12:32:11.968070 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15904dfe_b1ca_4922_95cf_6f3724fae181.slice/crio-36ae44d74e3cca090d25ba3ea797b71561f14710500966bbcb12af231e15c051 WatchSource:0}: Error finding container 36ae44d74e3cca090d25ba3ea797b71561f14710500966bbcb12af231e15c051: Status 404 returned error can't find the container with id 36ae44d74e3cca090d25ba3ea797b71561f14710500966bbcb12af231e15c051 Dec 15 12:32:12 crc kubenswrapper[4719]: I1215 12:32:12.263383 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" event={"ID":"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb","Type":"ContainerStarted","Data":"3d26ad91dc26542f020b8fde847eef7573baea280991fe4ff2ca57d62fe28232"} Dec 15 12:32:12 crc kubenswrapper[4719]: I1215 12:32:12.264767 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" event={"ID":"15904dfe-b1ca-4922-95cf-6f3724fae181","Type":"ContainerStarted","Data":"36ae44d74e3cca090d25ba3ea797b71561f14710500966bbcb12af231e15c051"} Dec 15 12:32:18 crc kubenswrapper[4719]: I1215 12:32:18.297357 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" event={"ID":"15904dfe-b1ca-4922-95cf-6f3724fae181","Type":"ContainerStarted","Data":"bbd35f25494360c027c0a61ad4fe0549c245539ca8df08b27de80ec9df125232"} Dec 15 12:32:18 crc kubenswrapper[4719]: I1215 12:32:18.297939 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:18 crc kubenswrapper[4719]: I1215 12:32:18.299915 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" event={"ID":"5009ea4d-5e13-4ac3-ad25-8c0976b0aafb","Type":"ContainerStarted","Data":"79f2ce3af9b60f10ea3407a63b87de17f536e51a49716a8afdd3167b30d59d97"} Dec 15 12:32:18 crc kubenswrapper[4719]: I1215 12:32:18.300113 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:18 crc kubenswrapper[4719]: I1215 12:32:18.329244 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" podStartSLOduration=1.8822890490000002 podStartE2EDuration="7.329224743s" podCreationTimestamp="2025-12-15 12:32:11 +0000 UTC" firstStartedPulling="2025-12-15 12:32:11.971293021 +0000 UTC m=+892.913586051" lastFinishedPulling="2025-12-15 12:32:17.418228715 +0000 UTC m=+898.360521745" observedRunningTime="2025-12-15 12:32:18.323888779 +0000 UTC m=+899.266181819" watchObservedRunningTime="2025-12-15 12:32:18.329224743 +0000 UTC m=+899.271517773" Dec 15 12:32:18 crc kubenswrapper[4719]: I1215 12:32:18.352774 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" podStartSLOduration=1.8331328679999999 podStartE2EDuration="7.352743097s" podCreationTimestamp="2025-12-15 12:32:11 +0000 UTC" firstStartedPulling="2025-12-15 12:32:11.884143111 +0000 UTC m=+892.826436141" lastFinishedPulling="2025-12-15 12:32:17.40375334 +0000 UTC m=+898.346046370" observedRunningTime="2025-12-15 12:32:18.351394875 +0000 UTC m=+899.293687905" watchObservedRunningTime="2025-12-15 12:32:18.352743097 +0000 UTC m=+899.295036127" Dec 15 12:32:21 crc kubenswrapper[4719]: I1215 12:32:21.356621 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:32:21 crc kubenswrapper[4719]: I1215 12:32:21.357009 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:32:31 crc kubenswrapper[4719]: I1215 12:32:31.730904 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-76fbf99b47-d626r" Dec 15 12:32:51 crc kubenswrapper[4719]: I1215 12:32:51.357187 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:32:51 crc kubenswrapper[4719]: I1215 12:32:51.357649 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:32:51 crc kubenswrapper[4719]: I1215 12:32:51.421913 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6cf5cc9b68-dpxck" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.104176 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-wnthh"] Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.107105 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.110469 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.110502 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.110691 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-lnppq" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.121281 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf"] Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.122112 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.124018 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.146073 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf"] Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.217725 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-tdxf5"] Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.218777 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.221469 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.221595 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.222896 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5dp2s" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.223730 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.247617 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5bddd4b946-96g4h"] Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.248465 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.252166 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.260137 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hktl7\" (UniqueName: \"kubernetes.io/projected/6bac73e3-46cd-4cca-bb26-1ae16f1c804e-kube-api-access-hktl7\") pod \"frr-k8s-webhook-server-7784b6fcf-crztf\" (UID: \"6bac73e3-46cd-4cca-bb26-1ae16f1c804e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.260306 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-metrics-certs\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.260404 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-frr-conf\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.260485 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49178e96-75a4-4b6d-8540-781182eb8123-metrics-certs\") pod \"controller-5bddd4b946-96g4h\" (UID: \"49178e96-75a4-4b6d-8540-781182eb8123\") " pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.260559 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-memberlist\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.260633 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-frr-sockets\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.260715 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-reloader\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.260784 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcgzw\" (UniqueName: \"kubernetes.io/projected/49178e96-75a4-4b6d-8540-781182eb8123-kube-api-access-lcgzw\") pod \"controller-5bddd4b946-96g4h\" (UID: \"49178e96-75a4-4b6d-8540-781182eb8123\") " pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.260873 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6prw\" (UniqueName: \"kubernetes.io/projected/dbea97f4-325a-4d31-8562-c1f5b546fca5-kube-api-access-r6prw\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.261009 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6bac73e3-46cd-4cca-bb26-1ae16f1c804e-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-crztf\" (UID: \"6bac73e3-46cd-4cca-bb26-1ae16f1c804e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.261096 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqk4k\" (UniqueName: \"kubernetes.io/projected/72e46a3d-9b79-404d-80ba-68a363c23b2b-kube-api-access-hqk4k\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.261187 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-metrics\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.261272 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/dbea97f4-325a-4d31-8562-c1f5b546fca5-frr-startup\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.261376 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/72e46a3d-9b79-404d-80ba-68a363c23b2b-metallb-excludel2\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.261455 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49178e96-75a4-4b6d-8540-781182eb8123-cert\") pod \"controller-5bddd4b946-96g4h\" (UID: \"49178e96-75a4-4b6d-8540-781182eb8123\") " pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.261539 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbea97f4-325a-4d31-8562-c1f5b546fca5-metrics-certs\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.266717 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-96g4h"] Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.362837 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/dbea97f4-325a-4d31-8562-c1f5b546fca5-frr-startup\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.362913 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49178e96-75a4-4b6d-8540-781182eb8123-cert\") pod \"controller-5bddd4b946-96g4h\" (UID: \"49178e96-75a4-4b6d-8540-781182eb8123\") " pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.362932 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/72e46a3d-9b79-404d-80ba-68a363c23b2b-metallb-excludel2\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.362981 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbea97f4-325a-4d31-8562-c1f5b546fca5-metrics-certs\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363009 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hktl7\" (UniqueName: \"kubernetes.io/projected/6bac73e3-46cd-4cca-bb26-1ae16f1c804e-kube-api-access-hktl7\") pod \"frr-k8s-webhook-server-7784b6fcf-crztf\" (UID: \"6bac73e3-46cd-4cca-bb26-1ae16f1c804e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363027 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-metrics-certs\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363044 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-frr-conf\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363064 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49178e96-75a4-4b6d-8540-781182eb8123-metrics-certs\") pod \"controller-5bddd4b946-96g4h\" (UID: \"49178e96-75a4-4b6d-8540-781182eb8123\") " pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363086 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-memberlist\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363101 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-frr-sockets\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363118 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-reloader\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: E1215 12:32:52.363128 4719 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 15 12:32:52 crc kubenswrapper[4719]: E1215 12:32:52.363192 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbea97f4-325a-4d31-8562-c1f5b546fca5-metrics-certs podName:dbea97f4-325a-4d31-8562-c1f5b546fca5 nodeName:}" failed. No retries permitted until 2025-12-15 12:32:52.863172716 +0000 UTC m=+933.805465746 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/dbea97f4-325a-4d31-8562-c1f5b546fca5-metrics-certs") pod "frr-k8s-wnthh" (UID: "dbea97f4-325a-4d31-8562-c1f5b546fca5") : secret "frr-k8s-certs-secret" not found Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363134 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcgzw\" (UniqueName: \"kubernetes.io/projected/49178e96-75a4-4b6d-8540-781182eb8123-kube-api-access-lcgzw\") pod \"controller-5bddd4b946-96g4h\" (UID: \"49178e96-75a4-4b6d-8540-781182eb8123\") " pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363314 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6prw\" (UniqueName: \"kubernetes.io/projected/dbea97f4-325a-4d31-8562-c1f5b546fca5-kube-api-access-r6prw\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: E1215 12:32:52.363326 4719 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 15 12:32:52 crc kubenswrapper[4719]: E1215 12:32:52.363361 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-memberlist podName:72e46a3d-9b79-404d-80ba-68a363c23b2b nodeName:}" failed. No retries permitted until 2025-12-15 12:32:52.863351922 +0000 UTC m=+933.805645012 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-memberlist") pod "speaker-tdxf5" (UID: "72e46a3d-9b79-404d-80ba-68a363c23b2b") : secret "metallb-memberlist" not found Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363379 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6bac73e3-46cd-4cca-bb26-1ae16f1c804e-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-crztf\" (UID: \"6bac73e3-46cd-4cca-bb26-1ae16f1c804e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363397 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqk4k\" (UniqueName: \"kubernetes.io/projected/72e46a3d-9b79-404d-80ba-68a363c23b2b-kube-api-access-hqk4k\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363430 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-metrics\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: E1215 12:32:52.363450 4719 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 15 12:32:52 crc kubenswrapper[4719]: E1215 12:32:52.363512 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49178e96-75a4-4b6d-8540-781182eb8123-metrics-certs podName:49178e96-75a4-4b6d-8540-781182eb8123 nodeName:}" failed. No retries permitted until 2025-12-15 12:32:52.863493706 +0000 UTC m=+933.805786786 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/49178e96-75a4-4b6d-8540-781182eb8123-metrics-certs") pod "controller-5bddd4b946-96g4h" (UID: "49178e96-75a4-4b6d-8540-781182eb8123") : secret "controller-certs-secret" not found Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363684 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-frr-conf\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.363784 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-metrics\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.364073 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-frr-sockets\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.364225 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/dbea97f4-325a-4d31-8562-c1f5b546fca5-reloader\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.364231 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/dbea97f4-325a-4d31-8562-c1f5b546fca5-frr-startup\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.364880 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/72e46a3d-9b79-404d-80ba-68a363c23b2b-metallb-excludel2\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.369498 4719 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.372654 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-metrics-certs\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.372718 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6bac73e3-46cd-4cca-bb26-1ae16f1c804e-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-crztf\" (UID: \"6bac73e3-46cd-4cca-bb26-1ae16f1c804e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.384394 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqk4k\" (UniqueName: \"kubernetes.io/projected/72e46a3d-9b79-404d-80ba-68a363c23b2b-kube-api-access-hqk4k\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.386644 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6prw\" (UniqueName: \"kubernetes.io/projected/dbea97f4-325a-4d31-8562-c1f5b546fca5-kube-api-access-r6prw\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.390056 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49178e96-75a4-4b6d-8540-781182eb8123-cert\") pod \"controller-5bddd4b946-96g4h\" (UID: \"49178e96-75a4-4b6d-8540-781182eb8123\") " pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.390909 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcgzw\" (UniqueName: \"kubernetes.io/projected/49178e96-75a4-4b6d-8540-781182eb8123-kube-api-access-lcgzw\") pod \"controller-5bddd4b946-96g4h\" (UID: \"49178e96-75a4-4b6d-8540-781182eb8123\") " pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.400704 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hktl7\" (UniqueName: \"kubernetes.io/projected/6bac73e3-46cd-4cca-bb26-1ae16f1c804e-kube-api-access-hktl7\") pod \"frr-k8s-webhook-server-7784b6fcf-crztf\" (UID: \"6bac73e3-46cd-4cca-bb26-1ae16f1c804e\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.443813 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.871561 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49178e96-75a4-4b6d-8540-781182eb8123-metrics-certs\") pod \"controller-5bddd4b946-96g4h\" (UID: \"49178e96-75a4-4b6d-8540-781182eb8123\") " pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.872005 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-memberlist\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.872101 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbea97f4-325a-4d31-8562-c1f5b546fca5-metrics-certs\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: E1215 12:32:52.872191 4719 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 15 12:32:52 crc kubenswrapper[4719]: E1215 12:32:52.872284 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-memberlist podName:72e46a3d-9b79-404d-80ba-68a363c23b2b nodeName:}" failed. No retries permitted until 2025-12-15 12:32:53.872263654 +0000 UTC m=+934.814556684 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-memberlist") pod "speaker-tdxf5" (UID: "72e46a3d-9b79-404d-80ba-68a363c23b2b") : secret "metallb-memberlist" not found Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.873878 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf"] Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.876920 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dbea97f4-325a-4d31-8562-c1f5b546fca5-metrics-certs\") pod \"frr-k8s-wnthh\" (UID: \"dbea97f4-325a-4d31-8562-c1f5b546fca5\") " pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:52 crc kubenswrapper[4719]: I1215 12:32:52.876982 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49178e96-75a4-4b6d-8540-781182eb8123-metrics-certs\") pod \"controller-5bddd4b946-96g4h\" (UID: \"49178e96-75a4-4b6d-8540-781182eb8123\") " pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:53 crc kubenswrapper[4719]: I1215 12:32:53.034397 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-wnthh" Dec 15 12:32:53 crc kubenswrapper[4719]: I1215 12:32:53.162730 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:53 crc kubenswrapper[4719]: I1215 12:32:53.563740 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-96g4h"] Dec 15 12:32:53 crc kubenswrapper[4719]: I1215 12:32:53.593953 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-96g4h" event={"ID":"49178e96-75a4-4b6d-8540-781182eb8123","Type":"ContainerStarted","Data":"8371bc321e727ca2ffa48f944d979ffe49708e77f797de0db005cfb96a0f340b"} Dec 15 12:32:53 crc kubenswrapper[4719]: I1215 12:32:53.594812 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" event={"ID":"6bac73e3-46cd-4cca-bb26-1ae16f1c804e","Type":"ContainerStarted","Data":"4a4dd86a3f0417af97898007089fb1a7d0985e4bcc229c5b66af984cccefb26c"} Dec 15 12:32:53 crc kubenswrapper[4719]: I1215 12:32:53.595828 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wnthh" event={"ID":"dbea97f4-325a-4d31-8562-c1f5b546fca5","Type":"ContainerStarted","Data":"cb6269d80660aed90f37d96e417fd630665c0c1dfdbfdb61c002dc852346d24a"} Dec 15 12:32:53 crc kubenswrapper[4719]: I1215 12:32:53.885126 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-memberlist\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:53 crc kubenswrapper[4719]: I1215 12:32:53.898572 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72e46a3d-9b79-404d-80ba-68a363c23b2b-memberlist\") pod \"speaker-tdxf5\" (UID: \"72e46a3d-9b79-404d-80ba-68a363c23b2b\") " pod="metallb-system/speaker-tdxf5" Dec 15 12:32:54 crc kubenswrapper[4719]: I1215 12:32:54.031374 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tdxf5" Dec 15 12:32:54 crc kubenswrapper[4719]: W1215 12:32:54.051253 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72e46a3d_9b79_404d_80ba_68a363c23b2b.slice/crio-261e307398ab343388a5b638c49e6d6be0900250b605abcd4800cfdf736628f9 WatchSource:0}: Error finding container 261e307398ab343388a5b638c49e6d6be0900250b605abcd4800cfdf736628f9: Status 404 returned error can't find the container with id 261e307398ab343388a5b638c49e6d6be0900250b605abcd4800cfdf736628f9 Dec 15 12:32:54 crc kubenswrapper[4719]: I1215 12:32:54.603111 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tdxf5" event={"ID":"72e46a3d-9b79-404d-80ba-68a363c23b2b","Type":"ContainerStarted","Data":"6bb8ae0a10acb998b97775e6ff2d7f86bcc5038c8aa71697b778460f8bbd450d"} Dec 15 12:32:54 crc kubenswrapper[4719]: I1215 12:32:54.603429 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tdxf5" event={"ID":"72e46a3d-9b79-404d-80ba-68a363c23b2b","Type":"ContainerStarted","Data":"261e307398ab343388a5b638c49e6d6be0900250b605abcd4800cfdf736628f9"} Dec 15 12:32:54 crc kubenswrapper[4719]: I1215 12:32:54.605394 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-96g4h" event={"ID":"49178e96-75a4-4b6d-8540-781182eb8123","Type":"ContainerStarted","Data":"33d9c0a6e63e738050e89f956514fe5d9132f079b6f3251f24191d6dda4e2555"} Dec 15 12:32:54 crc kubenswrapper[4719]: I1215 12:32:54.605429 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-96g4h" event={"ID":"49178e96-75a4-4b6d-8540-781182eb8123","Type":"ContainerStarted","Data":"47cacfda1d5333f78e86224df06934657c9ec588518b1c9412e1303c2b79764d"} Dec 15 12:32:54 crc kubenswrapper[4719]: I1215 12:32:54.605625 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:32:54 crc kubenswrapper[4719]: I1215 12:32:54.637639 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5bddd4b946-96g4h" podStartSLOduration=2.637618845 podStartE2EDuration="2.637618845s" podCreationTimestamp="2025-12-15 12:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:32:54.633463848 +0000 UTC m=+935.575756888" watchObservedRunningTime="2025-12-15 12:32:54.637618845 +0000 UTC m=+935.579911875" Dec 15 12:32:55 crc kubenswrapper[4719]: I1215 12:32:55.612979 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tdxf5" event={"ID":"72e46a3d-9b79-404d-80ba-68a363c23b2b","Type":"ContainerStarted","Data":"b4edaeb14ee16a1cd66c0916021df13b5c800193044a9e7ed56f3eec7f94ff4e"} Dec 15 12:32:55 crc kubenswrapper[4719]: I1215 12:32:55.647341 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-tdxf5" podStartSLOduration=3.6473199 podStartE2EDuration="3.6473199s" podCreationTimestamp="2025-12-15 12:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:32:55.642556244 +0000 UTC m=+936.584849274" watchObservedRunningTime="2025-12-15 12:32:55.6473199 +0000 UTC m=+936.589612940" Dec 15 12:32:56 crc kubenswrapper[4719]: I1215 12:32:56.620627 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-tdxf5" Dec 15 12:33:01 crc kubenswrapper[4719]: I1215 12:33:01.650052 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" event={"ID":"6bac73e3-46cd-4cca-bb26-1ae16f1c804e","Type":"ContainerStarted","Data":"b0f021becaaa68d744fd79b3c8f2fe12bdc05b9f3a354ebc44e34c69e9211c79"} Dec 15 12:33:01 crc kubenswrapper[4719]: I1215 12:33:01.650530 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" Dec 15 12:33:01 crc kubenswrapper[4719]: I1215 12:33:01.666250 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" podStartSLOduration=1.9209160490000001 podStartE2EDuration="9.666227637s" podCreationTimestamp="2025-12-15 12:32:52 +0000 UTC" firstStartedPulling="2025-12-15 12:32:52.878241256 +0000 UTC m=+933.820534286" lastFinishedPulling="2025-12-15 12:33:00.623552844 +0000 UTC m=+941.565845874" observedRunningTime="2025-12-15 12:33:01.66565134 +0000 UTC m=+942.607944370" watchObservedRunningTime="2025-12-15 12:33:01.666227637 +0000 UTC m=+942.608520667" Dec 15 12:33:02 crc kubenswrapper[4719]: I1215 12:33:02.658067 4719 generic.go:334] "Generic (PLEG): container finished" podID="dbea97f4-325a-4d31-8562-c1f5b546fca5" containerID="0805a5e4dca69f7063f377f5347d4bf53fee43208f366cf4f22d85d3c1901b40" exitCode=0 Dec 15 12:33:02 crc kubenswrapper[4719]: I1215 12:33:02.658171 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wnthh" event={"ID":"dbea97f4-325a-4d31-8562-c1f5b546fca5","Type":"ContainerDied","Data":"0805a5e4dca69f7063f377f5347d4bf53fee43208f366cf4f22d85d3c1901b40"} Dec 15 12:33:03 crc kubenswrapper[4719]: I1215 12:33:03.167436 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-96g4h" Dec 15 12:33:03 crc kubenswrapper[4719]: I1215 12:33:03.667660 4719 generic.go:334] "Generic (PLEG): container finished" podID="dbea97f4-325a-4d31-8562-c1f5b546fca5" containerID="f40ca5420ccf07d5c135e3cd7d6a665704d018d7324adbd79d81498b6e8e1662" exitCode=0 Dec 15 12:33:03 crc kubenswrapper[4719]: I1215 12:33:03.667697 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wnthh" event={"ID":"dbea97f4-325a-4d31-8562-c1f5b546fca5","Type":"ContainerDied","Data":"f40ca5420ccf07d5c135e3cd7d6a665704d018d7324adbd79d81498b6e8e1662"} Dec 15 12:33:04 crc kubenswrapper[4719]: I1215 12:33:04.035076 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-tdxf5" Dec 15 12:33:04 crc kubenswrapper[4719]: I1215 12:33:04.676588 4719 generic.go:334] "Generic (PLEG): container finished" podID="dbea97f4-325a-4d31-8562-c1f5b546fca5" containerID="cc6554a1bdc397ee89cdb8b83af1e729959071e0ba131efa20da39713aa8082d" exitCode=0 Dec 15 12:33:04 crc kubenswrapper[4719]: I1215 12:33:04.676702 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wnthh" event={"ID":"dbea97f4-325a-4d31-8562-c1f5b546fca5","Type":"ContainerDied","Data":"cc6554a1bdc397ee89cdb8b83af1e729959071e0ba131efa20da39713aa8082d"} Dec 15 12:33:05 crc kubenswrapper[4719]: I1215 12:33:05.687256 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wnthh" event={"ID":"dbea97f4-325a-4d31-8562-c1f5b546fca5","Type":"ContainerStarted","Data":"f8554b163e76d2b103998c6128c08148cfb18516b0e9b080549ec56cb0d33fdf"} Dec 15 12:33:05 crc kubenswrapper[4719]: I1215 12:33:05.687512 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wnthh" event={"ID":"dbea97f4-325a-4d31-8562-c1f5b546fca5","Type":"ContainerStarted","Data":"867b21c5373754dc1fe0c48b81f218642a6e03e94b05574151976f669e924547"} Dec 15 12:33:05 crc kubenswrapper[4719]: I1215 12:33:05.687523 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wnthh" event={"ID":"dbea97f4-325a-4d31-8562-c1f5b546fca5","Type":"ContainerStarted","Data":"2ca032d1416ddbb97f7d91dd12c8cdecd544589adeeb2dc85be8525d688d8031"} Dec 15 12:33:05 crc kubenswrapper[4719]: I1215 12:33:05.687532 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wnthh" event={"ID":"dbea97f4-325a-4d31-8562-c1f5b546fca5","Type":"ContainerStarted","Data":"effdcbb0847345ca5fbb0723d90a569398bc32de832dfe7c0865b5efda6613d2"} Dec 15 12:33:05 crc kubenswrapper[4719]: I1215 12:33:05.687540 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wnthh" event={"ID":"dbea97f4-325a-4d31-8562-c1f5b546fca5","Type":"ContainerStarted","Data":"344bb9aa7edaef4dc97503f91460913b51f618b49c619f220408c15f9856e54d"} Dec 15 12:33:06 crc kubenswrapper[4719]: I1215 12:33:06.702043 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-wnthh" event={"ID":"dbea97f4-325a-4d31-8562-c1f5b546fca5","Type":"ContainerStarted","Data":"7cc16a99fd3c0a5602d14385a50cd85bb2c86c23a3e68e8d65d4c169990e32a9"} Dec 15 12:33:06 crc kubenswrapper[4719]: I1215 12:33:06.734499 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-wnthh" podStartSLOduration=6.350445006 podStartE2EDuration="14.734477075s" podCreationTimestamp="2025-12-15 12:32:52 +0000 UTC" firstStartedPulling="2025-12-15 12:32:53.135512103 +0000 UTC m=+934.077805133" lastFinishedPulling="2025-12-15 12:33:01.519544162 +0000 UTC m=+942.461837202" observedRunningTime="2025-12-15 12:33:06.725551212 +0000 UTC m=+947.667844262" watchObservedRunningTime="2025-12-15 12:33:06.734477075 +0000 UTC m=+947.676770115" Dec 15 12:33:06 crc kubenswrapper[4719]: I1215 12:33:06.883897 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-lgbmr"] Dec 15 12:33:06 crc kubenswrapper[4719]: I1215 12:33:06.885406 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lgbmr" Dec 15 12:33:06 crc kubenswrapper[4719]: I1215 12:33:06.889278 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-wdmnt" Dec 15 12:33:06 crc kubenswrapper[4719]: I1215 12:33:06.889621 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 15 12:33:06 crc kubenswrapper[4719]: I1215 12:33:06.892500 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 15 12:33:06 crc kubenswrapper[4719]: I1215 12:33:06.900144 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lgbmr"] Dec 15 12:33:07 crc kubenswrapper[4719]: I1215 12:33:07.071712 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw57p\" (UniqueName: \"kubernetes.io/projected/ce469d36-18b5-41f5-a414-20696c9f44ef-kube-api-access-sw57p\") pod \"openstack-operator-index-lgbmr\" (UID: \"ce469d36-18b5-41f5-a414-20696c9f44ef\") " pod="openstack-operators/openstack-operator-index-lgbmr" Dec 15 12:33:07 crc kubenswrapper[4719]: I1215 12:33:07.172839 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw57p\" (UniqueName: \"kubernetes.io/projected/ce469d36-18b5-41f5-a414-20696c9f44ef-kube-api-access-sw57p\") pod \"openstack-operator-index-lgbmr\" (UID: \"ce469d36-18b5-41f5-a414-20696c9f44ef\") " pod="openstack-operators/openstack-operator-index-lgbmr" Dec 15 12:33:07 crc kubenswrapper[4719]: I1215 12:33:07.192551 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw57p\" (UniqueName: \"kubernetes.io/projected/ce469d36-18b5-41f5-a414-20696c9f44ef-kube-api-access-sw57p\") pod \"openstack-operator-index-lgbmr\" (UID: \"ce469d36-18b5-41f5-a414-20696c9f44ef\") " pod="openstack-operators/openstack-operator-index-lgbmr" Dec 15 12:33:07 crc kubenswrapper[4719]: I1215 12:33:07.203812 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lgbmr" Dec 15 12:33:07 crc kubenswrapper[4719]: I1215 12:33:07.626959 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lgbmr"] Dec 15 12:33:07 crc kubenswrapper[4719]: I1215 12:33:07.710549 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lgbmr" event={"ID":"ce469d36-18b5-41f5-a414-20696c9f44ef","Type":"ContainerStarted","Data":"80c267f906355dcd64f06af412369b90beb0ad938b07e287f141ca28a75a7e51"} Dec 15 12:33:07 crc kubenswrapper[4719]: I1215 12:33:07.710790 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-wnthh" Dec 15 12:33:08 crc kubenswrapper[4719]: I1215 12:33:08.035492 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-wnthh" Dec 15 12:33:08 crc kubenswrapper[4719]: I1215 12:33:08.082078 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-wnthh" Dec 15 12:33:10 crc kubenswrapper[4719]: I1215 12:33:10.260268 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-lgbmr"] Dec 15 12:33:10 crc kubenswrapper[4719]: I1215 12:33:10.737718 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lgbmr" event={"ID":"ce469d36-18b5-41f5-a414-20696c9f44ef","Type":"ContainerStarted","Data":"dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476"} Dec 15 12:33:10 crc kubenswrapper[4719]: I1215 12:33:10.737834 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-lgbmr" podUID="ce469d36-18b5-41f5-a414-20696c9f44ef" containerName="registry-server" containerID="cri-o://dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476" gracePeriod=2 Dec 15 12:33:10 crc kubenswrapper[4719]: I1215 12:33:10.757732 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-lgbmr" podStartSLOduration=2.099152034 podStartE2EDuration="4.757716919s" podCreationTimestamp="2025-12-15 12:33:06 +0000 UTC" firstStartedPulling="2025-12-15 12:33:07.640306213 +0000 UTC m=+948.582599243" lastFinishedPulling="2025-12-15 12:33:10.298871098 +0000 UTC m=+951.241164128" observedRunningTime="2025-12-15 12:33:10.752796628 +0000 UTC m=+951.695089668" watchObservedRunningTime="2025-12-15 12:33:10.757716919 +0000 UTC m=+951.700009949" Dec 15 12:33:10 crc kubenswrapper[4719]: I1215 12:33:10.881225 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zc4xm"] Dec 15 12:33:10 crc kubenswrapper[4719]: I1215 12:33:10.882196 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zc4xm" Dec 15 12:33:10 crc kubenswrapper[4719]: I1215 12:33:10.909730 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zc4xm"] Dec 15 12:33:10 crc kubenswrapper[4719]: I1215 12:33:10.922325 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh4qz\" (UniqueName: \"kubernetes.io/projected/bcaba66e-19b2-4c6f-9ec2-e545a7531624-kube-api-access-dh4qz\") pod \"openstack-operator-index-zc4xm\" (UID: \"bcaba66e-19b2-4c6f-9ec2-e545a7531624\") " pod="openstack-operators/openstack-operator-index-zc4xm" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.023976 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh4qz\" (UniqueName: \"kubernetes.io/projected/bcaba66e-19b2-4c6f-9ec2-e545a7531624-kube-api-access-dh4qz\") pod \"openstack-operator-index-zc4xm\" (UID: \"bcaba66e-19b2-4c6f-9ec2-e545a7531624\") " pod="openstack-operators/openstack-operator-index-zc4xm" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.052885 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh4qz\" (UniqueName: \"kubernetes.io/projected/bcaba66e-19b2-4c6f-9ec2-e545a7531624-kube-api-access-dh4qz\") pod \"openstack-operator-index-zc4xm\" (UID: \"bcaba66e-19b2-4c6f-9ec2-e545a7531624\") " pod="openstack-operators/openstack-operator-index-zc4xm" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.100918 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lgbmr" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.124618 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw57p\" (UniqueName: \"kubernetes.io/projected/ce469d36-18b5-41f5-a414-20696c9f44ef-kube-api-access-sw57p\") pod \"ce469d36-18b5-41f5-a414-20696c9f44ef\" (UID: \"ce469d36-18b5-41f5-a414-20696c9f44ef\") " Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.130076 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce469d36-18b5-41f5-a414-20696c9f44ef-kube-api-access-sw57p" (OuterVolumeSpecName: "kube-api-access-sw57p") pod "ce469d36-18b5-41f5-a414-20696c9f44ef" (UID: "ce469d36-18b5-41f5-a414-20696c9f44ef"). InnerVolumeSpecName "kube-api-access-sw57p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.202897 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zc4xm" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.225588 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw57p\" (UniqueName: \"kubernetes.io/projected/ce469d36-18b5-41f5-a414-20696c9f44ef-kube-api-access-sw57p\") on node \"crc\" DevicePath \"\"" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.584293 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zc4xm"] Dec 15 12:33:11 crc kubenswrapper[4719]: W1215 12:33:11.595292 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcaba66e_19b2_4c6f_9ec2_e545a7531624.slice/crio-4ea83ea224d0dc056af6e5a7720115f2258322fa394d5534fc3c7593ccd9dcf9 WatchSource:0}: Error finding container 4ea83ea224d0dc056af6e5a7720115f2258322fa394d5534fc3c7593ccd9dcf9: Status 404 returned error can't find the container with id 4ea83ea224d0dc056af6e5a7720115f2258322fa394d5534fc3c7593ccd9dcf9 Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.743005 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zc4xm" event={"ID":"bcaba66e-19b2-4c6f-9ec2-e545a7531624","Type":"ContainerStarted","Data":"4ea83ea224d0dc056af6e5a7720115f2258322fa394d5534fc3c7593ccd9dcf9"} Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.744569 4719 generic.go:334] "Generic (PLEG): container finished" podID="ce469d36-18b5-41f5-a414-20696c9f44ef" containerID="dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476" exitCode=0 Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.744605 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lgbmr" event={"ID":"ce469d36-18b5-41f5-a414-20696c9f44ef","Type":"ContainerDied","Data":"dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476"} Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.744620 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lgbmr" event={"ID":"ce469d36-18b5-41f5-a414-20696c9f44ef","Type":"ContainerDied","Data":"80c267f906355dcd64f06af412369b90beb0ad938b07e287f141ca28a75a7e51"} Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.744636 4719 scope.go:117] "RemoveContainer" containerID="dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.744668 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lgbmr" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.770379 4719 scope.go:117] "RemoveContainer" containerID="dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476" Dec 15 12:33:11 crc kubenswrapper[4719]: E1215 12:33:11.773901 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476\": container with ID starting with dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476 not found: ID does not exist" containerID="dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.773951 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476"} err="failed to get container status \"dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476\": rpc error: code = NotFound desc = could not find container \"dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476\": container with ID starting with dfaaf2b2d0e661c08cc8fd059c0b6d3e0d73d994fe89450ced14776cbb9a7476 not found: ID does not exist" Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.777116 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-lgbmr"] Dec 15 12:33:11 crc kubenswrapper[4719]: I1215 12:33:11.781896 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-lgbmr"] Dec 15 12:33:12 crc kubenswrapper[4719]: I1215 12:33:12.448417 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-crztf" Dec 15 12:33:12 crc kubenswrapper[4719]: I1215 12:33:12.751937 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zc4xm" event={"ID":"bcaba66e-19b2-4c6f-9ec2-e545a7531624","Type":"ContainerStarted","Data":"fefb5ab02d52357a5be45d030eb54f1f7f00b6ab70c92e8b93b3aaf1f615806d"} Dec 15 12:33:12 crc kubenswrapper[4719]: I1215 12:33:12.771352 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zc4xm" podStartSLOduration=2.714971839 podStartE2EDuration="2.771328822s" podCreationTimestamp="2025-12-15 12:33:10 +0000 UTC" firstStartedPulling="2025-12-15 12:33:11.601372636 +0000 UTC m=+952.543665676" lastFinishedPulling="2025-12-15 12:33:11.657729629 +0000 UTC m=+952.600022659" observedRunningTime="2025-12-15 12:33:12.770423024 +0000 UTC m=+953.712716054" watchObservedRunningTime="2025-12-15 12:33:12.771328822 +0000 UTC m=+953.713621852" Dec 15 12:33:13 crc kubenswrapper[4719]: I1215 12:33:13.484595 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce469d36-18b5-41f5-a414-20696c9f44ef" path="/var/lib/kubelet/pods/ce469d36-18b5-41f5-a414-20696c9f44ef/volumes" Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.203900 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zc4xm" Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.204439 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zc4xm" Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.229352 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zc4xm" Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.356400 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.356453 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.356491 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.356900 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d86014a7f0b55979e42be548b06831288da0c2f9ca42323013a65f25c4ad994"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.356960 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://2d86014a7f0b55979e42be548b06831288da0c2f9ca42323013a65f25c4ad994" gracePeriod=600 Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.804230 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="2d86014a7f0b55979e42be548b06831288da0c2f9ca42323013a65f25c4ad994" exitCode=0 Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.804389 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"2d86014a7f0b55979e42be548b06831288da0c2f9ca42323013a65f25c4ad994"} Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.804947 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"6512c03468894d57c0ec7081d538364858ba849e9d66b22fa695a25587406622"} Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.804970 4719 scope.go:117] "RemoveContainer" containerID="1ad2c6228b0260a3d44483166c7ea0f77e7d8e32151d0daaa5541fe5e5d67cc9" Dec 15 12:33:21 crc kubenswrapper[4719]: I1215 12:33:21.839993 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zc4xm" Dec 15 12:33:23 crc kubenswrapper[4719]: I1215 12:33:23.038166 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-wnthh" Dec 15 12:33:26 crc kubenswrapper[4719]: I1215 12:33:26.989515 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t"] Dec 15 12:33:26 crc kubenswrapper[4719]: E1215 12:33:26.990252 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce469d36-18b5-41f5-a414-20696c9f44ef" containerName="registry-server" Dec 15 12:33:26 crc kubenswrapper[4719]: I1215 12:33:26.990265 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce469d36-18b5-41f5-a414-20696c9f44ef" containerName="registry-server" Dec 15 12:33:26 crc kubenswrapper[4719]: I1215 12:33:26.990416 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce469d36-18b5-41f5-a414-20696c9f44ef" containerName="registry-server" Dec 15 12:33:26 crc kubenswrapper[4719]: I1215 12:33:26.991161 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:26 crc kubenswrapper[4719]: I1215 12:33:26.993065 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-2cn7s" Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.001705 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t"] Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.041370 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-bundle\") pod \"b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.041415 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8t4w\" (UniqueName: \"kubernetes.io/projected/c62f4038-b4ab-4211-8b78-e0ffe50814b7-kube-api-access-s8t4w\") pod \"b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.041451 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-util\") pod \"b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.142758 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-bundle\") pod \"b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.142809 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8t4w\" (UniqueName: \"kubernetes.io/projected/c62f4038-b4ab-4211-8b78-e0ffe50814b7-kube-api-access-s8t4w\") pod \"b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.142878 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-util\") pod \"b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.143254 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-bundle\") pod \"b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.143271 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-util\") pod \"b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.161787 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8t4w\" (UniqueName: \"kubernetes.io/projected/c62f4038-b4ab-4211-8b78-e0ffe50814b7-kube-api-access-s8t4w\") pod \"b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:27 crc kubenswrapper[4719]: I1215 12:33:27.308043 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:28 crc kubenswrapper[4719]: I1215 12:33:28.099165 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t"] Dec 15 12:33:28 crc kubenswrapper[4719]: I1215 12:33:28.843914 4719 generic.go:334] "Generic (PLEG): container finished" podID="c62f4038-b4ab-4211-8b78-e0ffe50814b7" containerID="b84538258d00f048ba9611cf571d5a7ced90d94e598ffb8fdf21205a68c89b05" exitCode=0 Dec 15 12:33:28 crc kubenswrapper[4719]: I1215 12:33:28.844002 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" event={"ID":"c62f4038-b4ab-4211-8b78-e0ffe50814b7","Type":"ContainerDied","Data":"b84538258d00f048ba9611cf571d5a7ced90d94e598ffb8fdf21205a68c89b05"} Dec 15 12:33:28 crc kubenswrapper[4719]: I1215 12:33:28.844135 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" event={"ID":"c62f4038-b4ab-4211-8b78-e0ffe50814b7","Type":"ContainerStarted","Data":"da8c3429247f2a00f57868dd01270c36a96e92092c87f664464ec294040166af"} Dec 15 12:33:29 crc kubenswrapper[4719]: I1215 12:33:29.852310 4719 generic.go:334] "Generic (PLEG): container finished" podID="c62f4038-b4ab-4211-8b78-e0ffe50814b7" containerID="6ac785b05c6ddee20bee14b346a9aa9b4c11df836773b6845671f9db97c579e2" exitCode=0 Dec 15 12:33:29 crc kubenswrapper[4719]: I1215 12:33:29.852361 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" event={"ID":"c62f4038-b4ab-4211-8b78-e0ffe50814b7","Type":"ContainerDied","Data":"6ac785b05c6ddee20bee14b346a9aa9b4c11df836773b6845671f9db97c579e2"} Dec 15 12:33:30 crc kubenswrapper[4719]: I1215 12:33:30.859840 4719 generic.go:334] "Generic (PLEG): container finished" podID="c62f4038-b4ab-4211-8b78-e0ffe50814b7" containerID="e3a20f36e9dc1634a0344cda474e13ee7d69aa21d4a9cdb8936c89585f204f14" exitCode=0 Dec 15 12:33:30 crc kubenswrapper[4719]: I1215 12:33:30.859972 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" event={"ID":"c62f4038-b4ab-4211-8b78-e0ffe50814b7","Type":"ContainerDied","Data":"e3a20f36e9dc1634a0344cda474e13ee7d69aa21d4a9cdb8936c89585f204f14"} Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.136404 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.309610 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-bundle\") pod \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.309693 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8t4w\" (UniqueName: \"kubernetes.io/projected/c62f4038-b4ab-4211-8b78-e0ffe50814b7-kube-api-access-s8t4w\") pod \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.309829 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-util\") pod \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\" (UID: \"c62f4038-b4ab-4211-8b78-e0ffe50814b7\") " Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.310549 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-bundle" (OuterVolumeSpecName: "bundle") pod "c62f4038-b4ab-4211-8b78-e0ffe50814b7" (UID: "c62f4038-b4ab-4211-8b78-e0ffe50814b7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.324956 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-util" (OuterVolumeSpecName: "util") pod "c62f4038-b4ab-4211-8b78-e0ffe50814b7" (UID: "c62f4038-b4ab-4211-8b78-e0ffe50814b7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.335101 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c62f4038-b4ab-4211-8b78-e0ffe50814b7-kube-api-access-s8t4w" (OuterVolumeSpecName: "kube-api-access-s8t4w") pod "c62f4038-b4ab-4211-8b78-e0ffe50814b7" (UID: "c62f4038-b4ab-4211-8b78-e0ffe50814b7"). InnerVolumeSpecName "kube-api-access-s8t4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.412149 4719 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-util\") on node \"crc\" DevicePath \"\"" Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.412220 4719 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c62f4038-b4ab-4211-8b78-e0ffe50814b7-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.412235 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8t4w\" (UniqueName: \"kubernetes.io/projected/c62f4038-b4ab-4211-8b78-e0ffe50814b7-kube-api-access-s8t4w\") on node \"crc\" DevicePath \"\"" Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.880214 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" event={"ID":"c62f4038-b4ab-4211-8b78-e0ffe50814b7","Type":"ContainerDied","Data":"da8c3429247f2a00f57868dd01270c36a96e92092c87f664464ec294040166af"} Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.880264 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da8c3429247f2a00f57868dd01270c36a96e92092c87f664464ec294040166af" Dec 15 12:33:32 crc kubenswrapper[4719]: I1215 12:33:32.880265 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.361668 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75"] Dec 15 12:33:39 crc kubenswrapper[4719]: E1215 12:33:39.362437 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62f4038-b4ab-4211-8b78-e0ffe50814b7" containerName="pull" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.362449 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62f4038-b4ab-4211-8b78-e0ffe50814b7" containerName="pull" Dec 15 12:33:39 crc kubenswrapper[4719]: E1215 12:33:39.362466 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62f4038-b4ab-4211-8b78-e0ffe50814b7" containerName="util" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.362471 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62f4038-b4ab-4211-8b78-e0ffe50814b7" containerName="util" Dec 15 12:33:39 crc kubenswrapper[4719]: E1215 12:33:39.362481 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62f4038-b4ab-4211-8b78-e0ffe50814b7" containerName="extract" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.362488 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62f4038-b4ab-4211-8b78-e0ffe50814b7" containerName="extract" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.362587 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c62f4038-b4ab-4211-8b78-e0ffe50814b7" containerName="extract" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.363051 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.366643 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-2qtpm" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.465335 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75"] Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.500949 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w247h\" (UniqueName: \"kubernetes.io/projected/131429f3-e15e-4107-af0e-b2d4c01f87af-kube-api-access-w247h\") pod \"openstack-operator-controller-operator-66c467f856-n9b75\" (UID: \"131429f3-e15e-4107-af0e-b2d4c01f87af\") " pod="openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.601596 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w247h\" (UniqueName: \"kubernetes.io/projected/131429f3-e15e-4107-af0e-b2d4c01f87af-kube-api-access-w247h\") pod \"openstack-operator-controller-operator-66c467f856-n9b75\" (UID: \"131429f3-e15e-4107-af0e-b2d4c01f87af\") " pod="openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.628937 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w247h\" (UniqueName: \"kubernetes.io/projected/131429f3-e15e-4107-af0e-b2d4c01f87af-kube-api-access-w247h\") pod \"openstack-operator-controller-operator-66c467f856-n9b75\" (UID: \"131429f3-e15e-4107-af0e-b2d4c01f87af\") " pod="openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.684335 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75" Dec 15 12:33:39 crc kubenswrapper[4719]: I1215 12:33:39.932326 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75"] Dec 15 12:33:40 crc kubenswrapper[4719]: I1215 12:33:40.934627 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75" event={"ID":"131429f3-e15e-4107-af0e-b2d4c01f87af","Type":"ContainerStarted","Data":"e576a790c5ca79c2b0c894de5132d4e82d7d07ef7245a34b42b8a8974150b2ea"} Dec 15 12:33:43 crc kubenswrapper[4719]: I1215 12:33:43.953473 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75" event={"ID":"131429f3-e15e-4107-af0e-b2d4c01f87af","Type":"ContainerStarted","Data":"93627f99e4688fe03faad9077d49a9fea19477f63610bd7abfacef126216dd7a"} Dec 15 12:33:43 crc kubenswrapper[4719]: I1215 12:33:43.954307 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75" Dec 15 12:33:43 crc kubenswrapper[4719]: I1215 12:33:43.986653 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75" podStartSLOduration=1.169644208 podStartE2EDuration="4.986636304s" podCreationTimestamp="2025-12-15 12:33:39 +0000 UTC" firstStartedPulling="2025-12-15 12:33:39.949209122 +0000 UTC m=+980.891502152" lastFinishedPulling="2025-12-15 12:33:43.766201208 +0000 UTC m=+984.708494248" observedRunningTime="2025-12-15 12:33:43.98099662 +0000 UTC m=+984.923289680" watchObservedRunningTime="2025-12-15 12:33:43.986636304 +0000 UTC m=+984.928929334" Dec 15 12:33:49 crc kubenswrapper[4719]: I1215 12:33:49.687413 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-66c467f856-n9b75" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.421697 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-95949466-k2fz4"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.423067 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-95949466-k2fz4" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.434969 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-9nxd2" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.440204 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-95949466-k2fz4"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.444060 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.444772 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.449308 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-66xvz" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.460687 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.468182 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.469164 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.471493 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-r7tmf" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.489044 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-767f9d7567-2866h"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.489939 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" Dec 15 12:34:06 crc kubenswrapper[4719]: W1215 12:34:06.495874 4719 reflector.go:561] object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-sqvwf": failed to list *v1.Secret: secrets "glance-operator-controller-manager-dockercfg-sqvwf" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Dec 15 12:34:06 crc kubenswrapper[4719]: E1215 12:34:06.495916 4719 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"glance-operator-controller-manager-dockercfg-sqvwf\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"glance-operator-controller-manager-dockercfg-sqvwf\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.504430 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.511430 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-767f9d7567-2866h"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.567627 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.568651 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.571336 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-zfpk6" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.580559 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n65v2\" (UniqueName: \"kubernetes.io/projected/7f47c007-e5bf-47dd-8905-69ed7cfc3a9a-kube-api-access-n65v2\") pod \"cinder-operator-controller-manager-5cf45c46bd-pmmwk\" (UID: \"7f47c007-e5bf-47dd-8905-69ed7cfc3a9a\") " pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.580611 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjq7c\" (UniqueName: \"kubernetes.io/projected/5c7fa683-b59c-45ef-bc29-92400f43e4b8-kube-api-access-bjq7c\") pod \"barbican-operator-controller-manager-95949466-k2fz4\" (UID: \"5c7fa683-b59c-45ef-bc29-92400f43e4b8\") " pod="openstack-operators/barbican-operator-controller-manager-95949466-k2fz4" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.580718 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7grb4\" (UniqueName: \"kubernetes.io/projected/7ced32f3-ca3c-406c-966b-93be87cd6d25-kube-api-access-7grb4\") pod \"designate-operator-controller-manager-66f8b87655-5td8v\" (UID: \"7ced32f3-ca3c-406c-966b-93be87cd6d25\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.586631 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.605894 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.607155 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.610752 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-2vjjv" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.619815 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.630186 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.631374 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.650223 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.651052 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.654645 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.659437 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-h4nt9" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.660040 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-xtll6" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.674030 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.678420 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.688271 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6cgj\" (UniqueName: \"kubernetes.io/projected/e6adc1e0-5c73-431d-a94d-e2eabc16dd03-kube-api-access-n6cgj\") pod \"ironic-operator-controller-manager-f458558d7-zkv6s\" (UID: \"e6adc1e0-5c73-431d-a94d-e2eabc16dd03\") " pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.688561 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq2n9\" (UniqueName: \"kubernetes.io/projected/345b77d5-0419-433e-b7d4-053a0b6fcf86-kube-api-access-vq2n9\") pod \"glance-operator-controller-manager-767f9d7567-2866h\" (UID: \"345b77d5-0419-433e-b7d4-053a0b6fcf86\") " pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.688693 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw798\" (UniqueName: \"kubernetes.io/projected/b6efe10d-9999-4340-822b-48fb8fb99f14-kube-api-access-mw798\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.688806 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.688955 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgmf5\" (UniqueName: \"kubernetes.io/projected/6cdd3328-cd4d-4e90-8ab9-d0953cab73de-kube-api-access-pgmf5\") pod \"heat-operator-controller-manager-59b8dcb766-llbls\" (UID: \"6cdd3328-cd4d-4e90-8ab9-d0953cab73de\") " pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.689084 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7grb4\" (UniqueName: \"kubernetes.io/projected/7ced32f3-ca3c-406c-966b-93be87cd6d25-kube-api-access-7grb4\") pod \"designate-operator-controller-manager-66f8b87655-5td8v\" (UID: \"7ced32f3-ca3c-406c-966b-93be87cd6d25\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.689208 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n65v2\" (UniqueName: \"kubernetes.io/projected/7f47c007-e5bf-47dd-8905-69ed7cfc3a9a-kube-api-access-n65v2\") pod \"cinder-operator-controller-manager-5cf45c46bd-pmmwk\" (UID: \"7f47c007-e5bf-47dd-8905-69ed7cfc3a9a\") " pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.699351 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc9q8\" (UniqueName: \"kubernetes.io/projected/15643ba0-8573-4f51-8ee6-9fd78e10b6a0-kube-api-access-qc9q8\") pod \"horizon-operator-controller-manager-6ccf486b9-scntm\" (UID: \"15643ba0-8573-4f51-8ee6-9fd78e10b6a0\") " pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.699595 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjq7c\" (UniqueName: \"kubernetes.io/projected/5c7fa683-b59c-45ef-bc29-92400f43e4b8-kube-api-access-bjq7c\") pod \"barbican-operator-controller-manager-95949466-k2fz4\" (UID: \"5c7fa683-b59c-45ef-bc29-92400f43e4b8\") " pod="openstack-operators/barbican-operator-controller-manager-95949466-k2fz4" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.699393 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.713673 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.715629 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-td5hr" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.745971 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7grb4\" (UniqueName: \"kubernetes.io/projected/7ced32f3-ca3c-406c-966b-93be87cd6d25-kube-api-access-7grb4\") pod \"designate-operator-controller-manager-66f8b87655-5td8v\" (UID: \"7ced32f3-ca3c-406c-966b-93be87cd6d25\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.751393 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n65v2\" (UniqueName: \"kubernetes.io/projected/7f47c007-e5bf-47dd-8905-69ed7cfc3a9a-kube-api-access-n65v2\") pod \"cinder-operator-controller-manager-5cf45c46bd-pmmwk\" (UID: \"7f47c007-e5bf-47dd-8905-69ed7cfc3a9a\") " pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.760486 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.762038 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.785110 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjq7c\" (UniqueName: \"kubernetes.io/projected/5c7fa683-b59c-45ef-bc29-92400f43e4b8-kube-api-access-bjq7c\") pod \"barbican-operator-controller-manager-95949466-k2fz4\" (UID: \"5c7fa683-b59c-45ef-bc29-92400f43e4b8\") " pod="openstack-operators/barbican-operator-controller-manager-95949466-k2fz4" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.791796 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.792465 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.792766 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.799682 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-l6hzv" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.803292 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq2n9\" (UniqueName: \"kubernetes.io/projected/345b77d5-0419-433e-b7d4-053a0b6fcf86-kube-api-access-vq2n9\") pod \"glance-operator-controller-manager-767f9d7567-2866h\" (UID: \"345b77d5-0419-433e-b7d4-053a0b6fcf86\") " pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.803340 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6cgj\" (UniqueName: \"kubernetes.io/projected/e6adc1e0-5c73-431d-a94d-e2eabc16dd03-kube-api-access-n6cgj\") pod \"ironic-operator-controller-manager-f458558d7-zkv6s\" (UID: \"e6adc1e0-5c73-431d-a94d-e2eabc16dd03\") " pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.803377 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw798\" (UniqueName: \"kubernetes.io/projected/b6efe10d-9999-4340-822b-48fb8fb99f14-kube-api-access-mw798\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.803410 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwncv\" (UniqueName: \"kubernetes.io/projected/571dd62a-58e8-4dab-ad04-a95621a65078-kube-api-access-pwncv\") pod \"manila-operator-controller-manager-5fdd9786f7-89zrh\" (UID: \"571dd62a-58e8-4dab-ad04-a95621a65078\") " pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.803432 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.803458 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgmf5\" (UniqueName: \"kubernetes.io/projected/6cdd3328-cd4d-4e90-8ab9-d0953cab73de-kube-api-access-pgmf5\") pod \"heat-operator-controller-manager-59b8dcb766-llbls\" (UID: \"6cdd3328-cd4d-4e90-8ab9-d0953cab73de\") " pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.803510 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc9q8\" (UniqueName: \"kubernetes.io/projected/15643ba0-8573-4f51-8ee6-9fd78e10b6a0-kube-api-access-qc9q8\") pod \"horizon-operator-controller-manager-6ccf486b9-scntm\" (UID: \"15643ba0-8573-4f51-8ee6-9fd78e10b6a0\") " pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" Dec 15 12:34:06 crc kubenswrapper[4719]: E1215 12:34:06.804278 4719 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:06 crc kubenswrapper[4719]: E1215 12:34:06.804325 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert podName:b6efe10d-9999-4340-822b-48fb8fb99f14 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:07.304309455 +0000 UTC m=+1008.246602485 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert") pod "infra-operator-controller-manager-58944d7758-wh4dh" (UID: "b6efe10d-9999-4340-822b-48fb8fb99f14") : secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.845954 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.846507 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6cgj\" (UniqueName: \"kubernetes.io/projected/e6adc1e0-5c73-431d-a94d-e2eabc16dd03-kube-api-access-n6cgj\") pod \"ironic-operator-controller-manager-f458558d7-zkv6s\" (UID: \"e6adc1e0-5c73-431d-a94d-e2eabc16dd03\") " pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.848720 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.848884 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.860434 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq2n9\" (UniqueName: \"kubernetes.io/projected/345b77d5-0419-433e-b7d4-053a0b6fcf86-kube-api-access-vq2n9\") pod \"glance-operator-controller-manager-767f9d7567-2866h\" (UID: \"345b77d5-0419-433e-b7d4-053a0b6fcf86\") " pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.879986 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgmf5\" (UniqueName: \"kubernetes.io/projected/6cdd3328-cd4d-4e90-8ab9-d0953cab73de-kube-api-access-pgmf5\") pod \"heat-operator-controller-manager-59b8dcb766-llbls\" (UID: \"6cdd3328-cd4d-4e90-8ab9-d0953cab73de\") " pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.880317 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.880504 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-fjlq7" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.880900 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc9q8\" (UniqueName: \"kubernetes.io/projected/15643ba0-8573-4f51-8ee6-9fd78e10b6a0-kube-api-access-qc9q8\") pod \"horizon-operator-controller-manager-6ccf486b9-scntm\" (UID: \"15643ba0-8573-4f51-8ee6-9fd78e10b6a0\") " pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.882397 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw798\" (UniqueName: \"kubernetes.io/projected/b6efe10d-9999-4340-822b-48fb8fb99f14-kube-api-access-mw798\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.885177 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.886871 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.890043 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-hhrlj" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.906018 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.907206 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hjpg\" (UniqueName: \"kubernetes.io/projected/def1648e-ccf3-4a95-9a44-92bf0f83cb49-kube-api-access-7hjpg\") pod \"mariadb-operator-controller-manager-f76f4954c-zhvw6\" (UID: \"def1648e-ccf3-4a95-9a44-92bf0f83cb49\") " pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.907267 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm699\" (UniqueName: \"kubernetes.io/projected/83a0cff1-8c43-4bb1-aa4f-66c4662fd235-kube-api-access-qm699\") pod \"neutron-operator-controller-manager-7cd87b778f-hb66s\" (UID: \"83a0cff1-8c43-4bb1-aa4f-66c4662fd235\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.907293 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwncv\" (UniqueName: \"kubernetes.io/projected/571dd62a-58e8-4dab-ad04-a95621a65078-kube-api-access-pwncv\") pod \"manila-operator-controller-manager-5fdd9786f7-89zrh\" (UID: \"571dd62a-58e8-4dab-ad04-a95621a65078\") " pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.919943 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.920835 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.921417 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.921734 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.926777 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.929762 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-l48jn" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.929807 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-glx6q" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.930171 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.938190 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.942144 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.966528 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556"] Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.967568 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.972098 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8f4t2" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.972262 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 15 12:34:06 crc kubenswrapper[4719]: I1215 12:34:06.984589 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwncv\" (UniqueName: \"kubernetes.io/projected/571dd62a-58e8-4dab-ad04-a95621a65078-kube-api-access-pwncv\") pod \"manila-operator-controller-manager-5fdd9786f7-89zrh\" (UID: \"571dd62a-58e8-4dab-ad04-a95621a65078\") " pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.003324 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.009904 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hjpg\" (UniqueName: \"kubernetes.io/projected/def1648e-ccf3-4a95-9a44-92bf0f83cb49-kube-api-access-7hjpg\") pod \"mariadb-operator-controller-manager-f76f4954c-zhvw6\" (UID: \"def1648e-ccf3-4a95-9a44-92bf0f83cb49\") " pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.009987 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btwv6\" (UniqueName: \"kubernetes.io/projected/9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc-kube-api-access-btwv6\") pod \"keystone-operator-controller-manager-5c7cbf548f-5slz5\" (UID: \"9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc\") " pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.010032 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67f9w\" (UniqueName: \"kubernetes.io/projected/3b66851a-75d1-46ea-8d06-ad7385fdff6b-kube-api-access-67f9w\") pod \"octavia-operator-controller-manager-68c649d9d-p7z6f\" (UID: \"3b66851a-75d1-46ea-8d06-ad7385fdff6b\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.010052 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm699\" (UniqueName: \"kubernetes.io/projected/83a0cff1-8c43-4bb1-aa4f-66c4662fd235-kube-api-access-qm699\") pod \"neutron-operator-controller-manager-7cd87b778f-hb66s\" (UID: \"83a0cff1-8c43-4bb1-aa4f-66c4662fd235\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.010298 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghmnd\" (UniqueName: \"kubernetes.io/projected/828d7a4d-0956-40fc-967f-e52d1701953c-kube-api-access-ghmnd\") pod \"nova-operator-controller-manager-5fbbf8b6cc-c9szj\" (UID: \"828d7a4d-0956-40fc-967f-e52d1701953c\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.010315 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.010338 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4mfg\" (UniqueName: \"kubernetes.io/projected/24c5b173-c013-45e1-9563-1e49556b23a5-kube-api-access-k4mfg\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.028667 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.039904 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.045134 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.057286 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.058301 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.064440 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-8tglg" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.065514 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-95949466-k2fz4" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.068408 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm699\" (UniqueName: \"kubernetes.io/projected/83a0cff1-8c43-4bb1-aa4f-66c4662fd235-kube-api-access-qm699\") pod \"neutron-operator-controller-manager-7cd87b778f-hb66s\" (UID: \"83a0cff1-8c43-4bb1-aa4f-66c4662fd235\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.071313 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-ls6f9" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.094721 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.111691 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btwv6\" (UniqueName: \"kubernetes.io/projected/9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc-kube-api-access-btwv6\") pod \"keystone-operator-controller-manager-5c7cbf548f-5slz5\" (UID: \"9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc\") " pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.111750 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67f9w\" (UniqueName: \"kubernetes.io/projected/3b66851a-75d1-46ea-8d06-ad7385fdff6b-kube-api-access-67f9w\") pod \"octavia-operator-controller-manager-68c649d9d-p7z6f\" (UID: \"3b66851a-75d1-46ea-8d06-ad7385fdff6b\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.111788 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghmnd\" (UniqueName: \"kubernetes.io/projected/828d7a4d-0956-40fc-967f-e52d1701953c-kube-api-access-ghmnd\") pod \"nova-operator-controller-manager-5fbbf8b6cc-c9szj\" (UID: \"828d7a4d-0956-40fc-967f-e52d1701953c\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.111808 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.111836 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4mfg\" (UniqueName: \"kubernetes.io/projected/24c5b173-c013-45e1-9563-1e49556b23a5-kube-api-access-k4mfg\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:07 crc kubenswrapper[4719]: E1215 12:34:07.113117 4719 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:07 crc kubenswrapper[4719]: E1215 12:34:07.113167 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert podName:24c5b173-c013-45e1-9563-1e49556b23a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:07.613151984 +0000 UTC m=+1008.555445014 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" (UID: "24c5b173-c013-45e1-9563-1e49556b23a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.141566 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hjpg\" (UniqueName: \"kubernetes.io/projected/def1648e-ccf3-4a95-9a44-92bf0f83cb49-kube-api-access-7hjpg\") pod \"mariadb-operator-controller-manager-f76f4954c-zhvw6\" (UID: \"def1648e-ccf3-4a95-9a44-92bf0f83cb49\") " pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.157293 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghmnd\" (UniqueName: \"kubernetes.io/projected/828d7a4d-0956-40fc-967f-e52d1701953c-kube-api-access-ghmnd\") pod \"nova-operator-controller-manager-5fbbf8b6cc-c9szj\" (UID: \"828d7a4d-0956-40fc-967f-e52d1701953c\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.172425 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.180311 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67f9w\" (UniqueName: \"kubernetes.io/projected/3b66851a-75d1-46ea-8d06-ad7385fdff6b-kube-api-access-67f9w\") pod \"octavia-operator-controller-manager-68c649d9d-p7z6f\" (UID: \"3b66851a-75d1-46ea-8d06-ad7385fdff6b\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.188126 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btwv6\" (UniqueName: \"kubernetes.io/projected/9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc-kube-api-access-btwv6\") pod \"keystone-operator-controller-manager-5c7cbf548f-5slz5\" (UID: \"9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc\") " pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.210898 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.216168 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk7gs\" (UniqueName: \"kubernetes.io/projected/d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e-kube-api-access-vk7gs\") pod \"ovn-operator-controller-manager-bf6d4f946-gn5n9\" (UID: \"d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.216207 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw2w6\" (UniqueName: \"kubernetes.io/projected/9adcd81c-e72b-44b8-bd4f-36e49c8e5c82-kube-api-access-bw2w6\") pod \"placement-operator-controller-manager-8665b56d78-wslbg\" (UID: \"9adcd81c-e72b-44b8-bd4f-36e49c8e5c82\") " pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.216980 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.243528 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4mfg\" (UniqueName: \"kubernetes.io/projected/24c5b173-c013-45e1-9563-1e49556b23a5-kube-api-access-k4mfg\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.253420 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.254663 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.261602 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.268277 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-6tf6q" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.268780 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.311489 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.312763 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.315003 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-cx6gs" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.319076 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk7gs\" (UniqueName: \"kubernetes.io/projected/d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e-kube-api-access-vk7gs\") pod \"ovn-operator-controller-manager-bf6d4f946-gn5n9\" (UID: \"d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.319116 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw2w6\" (UniqueName: \"kubernetes.io/projected/9adcd81c-e72b-44b8-bd4f-36e49c8e5c82-kube-api-access-bw2w6\") pod \"placement-operator-controller-manager-8665b56d78-wslbg\" (UID: \"9adcd81c-e72b-44b8-bd4f-36e49c8e5c82\") " pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.319142 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnrcb\" (UniqueName: \"kubernetes.io/projected/c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157-kube-api-access-mnrcb\") pod \"swift-operator-controller-manager-5c6df8f9-d9d67\" (UID: \"c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157\") " pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.319186 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfq4x\" (UniqueName: \"kubernetes.io/projected/375826ba-0785-4935-96c2-a2e5345aaf45-kube-api-access-kfq4x\") pod \"telemetry-operator-controller-manager-97d456b9-qxhfl\" (UID: \"375826ba-0785-4935-96c2-a2e5345aaf45\") " pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.319209 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:07 crc kubenswrapper[4719]: E1215 12:34:07.319323 4719 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:07 crc kubenswrapper[4719]: E1215 12:34:07.319363 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert podName:b6efe10d-9999-4340-822b-48fb8fb99f14 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:08.319349049 +0000 UTC m=+1009.261642079 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert") pod "infra-operator-controller-manager-58944d7758-wh4dh" (UID: "b6efe10d-9999-4340-822b-48fb8fb99f14") : secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.320771 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.332445 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.346198 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.376343 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.377467 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.378194 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk7gs\" (UniqueName: \"kubernetes.io/projected/d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e-kube-api-access-vk7gs\") pod \"ovn-operator-controller-manager-bf6d4f946-gn5n9\" (UID: \"d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.380454 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw2w6\" (UniqueName: \"kubernetes.io/projected/9adcd81c-e72b-44b8-bd4f-36e49c8e5c82-kube-api-access-bw2w6\") pod \"placement-operator-controller-manager-8665b56d78-wslbg\" (UID: \"9adcd81c-e72b-44b8-bd4f-36e49c8e5c82\") " pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.384014 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-xkj9f" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.384054 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.384467 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.396919 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.397869 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.403591 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-qzq8k" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.438397 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnrcb\" (UniqueName: \"kubernetes.io/projected/c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157-kube-api-access-mnrcb\") pod \"swift-operator-controller-manager-5c6df8f9-d9d67\" (UID: \"c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157\") " pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.438565 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfq4x\" (UniqueName: \"kubernetes.io/projected/375826ba-0785-4935-96c2-a2e5345aaf45-kube-api-access-kfq4x\") pod \"telemetry-operator-controller-manager-97d456b9-qxhfl\" (UID: \"375826ba-0785-4935-96c2-a2e5345aaf45\") " pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.464988 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfq4x\" (UniqueName: \"kubernetes.io/projected/375826ba-0785-4935-96c2-a2e5345aaf45-kube-api-access-kfq4x\") pod \"telemetry-operator-controller-manager-97d456b9-qxhfl\" (UID: \"375826ba-0785-4935-96c2-a2e5345aaf45\") " pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.473664 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.478760 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnrcb\" (UniqueName: \"kubernetes.io/projected/c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157-kube-api-access-mnrcb\") pod \"swift-operator-controller-manager-5c6df8f9-d9d67\" (UID: \"c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157\") " pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.488196 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.540549 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmgrr\" (UniqueName: \"kubernetes.io/projected/a9ebdf36-2095-47b8-8edf-89036ffdee84-kube-api-access-qmgrr\") pod \"watcher-operator-controller-manager-55f78b7c4c-292vp\" (UID: \"a9ebdf36-2095-47b8-8edf-89036ffdee84\") " pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.541270 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmcmd\" (UniqueName: \"kubernetes.io/projected/72cd4bb7-5783-492b-bae4-1a9fa633cd7b-kube-api-access-vmcmd\") pod \"test-operator-controller-manager-756ccf86c7-l98v4\" (UID: \"72cd4bb7-5783-492b-bae4-1a9fa633cd7b\") " pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.574392 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.592263 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.593472 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.596816 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.597696 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.598102 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-kfj4g" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.646712 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmgrr\" (UniqueName: \"kubernetes.io/projected/a9ebdf36-2095-47b8-8edf-89036ffdee84-kube-api-access-qmgrr\") pod \"watcher-operator-controller-manager-55f78b7c4c-292vp\" (UID: \"a9ebdf36-2095-47b8-8edf-89036ffdee84\") " pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.647031 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.647105 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.647154 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.647174 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmcmd\" (UniqueName: \"kubernetes.io/projected/72cd4bb7-5783-492b-bae4-1a9fa633cd7b-kube-api-access-vmcmd\") pod \"test-operator-controller-manager-756ccf86c7-l98v4\" (UID: \"72cd4bb7-5783-492b-bae4-1a9fa633cd7b\") " pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.647195 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxfhq\" (UniqueName: \"kubernetes.io/projected/db19bab3-733c-4ef1-8f46-df4cf5842701-kube-api-access-jxfhq\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:07 crc kubenswrapper[4719]: E1215 12:34:07.647905 4719 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:07 crc kubenswrapper[4719]: E1215 12:34:07.647959 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert podName:24c5b173-c013-45e1-9563-1e49556b23a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:08.647931818 +0000 UTC m=+1009.590224848 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" (UID: "24c5b173-c013-45e1-9563-1e49556b23a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.674802 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.680325 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.689532 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.692122 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmcmd\" (UniqueName: \"kubernetes.io/projected/72cd4bb7-5783-492b-bae4-1a9fa633cd7b-kube-api-access-vmcmd\") pod \"test-operator-controller-manager-756ccf86c7-l98v4\" (UID: \"72cd4bb7-5783-492b-bae4-1a9fa633cd7b\") " pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.693452 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmgrr\" (UniqueName: \"kubernetes.io/projected/a9ebdf36-2095-47b8-8edf-89036ffdee84-kube-api-access-qmgrr\") pod \"watcher-operator-controller-manager-55f78b7c4c-292vp\" (UID: \"a9ebdf36-2095-47b8-8edf-89036ffdee84\") " pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.695841 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.696732 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.698480 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.713910 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-ztmw2" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.735707 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.739663 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.748724 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.748768 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxfhq\" (UniqueName: \"kubernetes.io/projected/db19bab3-733c-4ef1-8f46-df4cf5842701-kube-api-access-jxfhq\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.748837 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:07 crc kubenswrapper[4719]: E1215 12:34:07.749000 4719 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 15 12:34:07 crc kubenswrapper[4719]: E1215 12:34:07.749049 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:08.249032985 +0000 UTC m=+1009.191326015 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "metrics-server-cert" not found Dec 15 12:34:07 crc kubenswrapper[4719]: E1215 12:34:07.749842 4719 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 15 12:34:07 crc kubenswrapper[4719]: E1215 12:34:07.749891 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:08.24988275 +0000 UTC m=+1009.192175780 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "webhook-server-cert" not found Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.774441 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxfhq\" (UniqueName: \"kubernetes.io/projected/db19bab3-733c-4ef1-8f46-df4cf5842701-kube-api-access-jxfhq\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.844241 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.853923 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzwgw\" (UniqueName: \"kubernetes.io/projected/0c36396e-131f-4154-b82e-490b36d7fa63-kube-api-access-wzwgw\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jzg2t\" (UID: \"0c36396e-131f-4154-b82e-490b36d7fa63\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.863101 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.871581 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk"] Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.913750 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-sqvwf" Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.918345 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" Dec 15 12:34:07 crc kubenswrapper[4719]: W1215 12:34:07.931874 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ced32f3_ca3c_406c_966b_93be87cd6d25.slice/crio-239d624c6f54ece21d1893341390666e34c8e417a3ab7dfd3c5f2609a10b6203 WatchSource:0}: Error finding container 239d624c6f54ece21d1893341390666e34c8e417a3ab7dfd3c5f2609a10b6203: Status 404 returned error can't find the container with id 239d624c6f54ece21d1893341390666e34c8e417a3ab7dfd3c5f2609a10b6203 Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.958672 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzwgw\" (UniqueName: \"kubernetes.io/projected/0c36396e-131f-4154-b82e-490b36d7fa63-kube-api-access-wzwgw\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jzg2t\" (UID: \"0c36396e-131f-4154-b82e-490b36d7fa63\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" Dec 15 12:34:07 crc kubenswrapper[4719]: W1215 12:34:07.979068 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f47c007_e5bf_47dd_8905_69ed7cfc3a9a.slice/crio-cbd753b9137d75fef555a97de1c2f85f6b2939d18e4e2fd45527a86a18e38d79 WatchSource:0}: Error finding container cbd753b9137d75fef555a97de1c2f85f6b2939d18e4e2fd45527a86a18e38d79: Status 404 returned error can't find the container with id cbd753b9137d75fef555a97de1c2f85f6b2939d18e4e2fd45527a86a18e38d79 Dec 15 12:34:07 crc kubenswrapper[4719]: I1215 12:34:07.990456 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzwgw\" (UniqueName: \"kubernetes.io/projected/0c36396e-131f-4154-b82e-490b36d7fa63-kube-api-access-wzwgw\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jzg2t\" (UID: \"0c36396e-131f-4154-b82e-490b36d7fa63\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.044700 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm"] Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.118448 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" event={"ID":"7f47c007-e5bf-47dd-8905-69ed7cfc3a9a","Type":"ContainerStarted","Data":"cbd753b9137d75fef555a97de1c2f85f6b2939d18e4e2fd45527a86a18e38d79"} Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.121879 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" event={"ID":"6cdd3328-cd4d-4e90-8ab9-d0953cab73de","Type":"ContainerStarted","Data":"8816244c894a6edcbe82a9155488f2b42ec18f7fdaa0f5f3b573c8d8d099d293"} Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.124187 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" event={"ID":"15643ba0-8573-4f51-8ee6-9fd78e10b6a0","Type":"ContainerStarted","Data":"f84b0f07cbc0291f2fa81387595a7c1ae2a7bd68767792ce28d23a435eae83c3"} Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.135381 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v" event={"ID":"7ced32f3-ca3c-406c-966b-93be87cd6d25","Type":"ContainerStarted","Data":"239d624c6f54ece21d1893341390666e34c8e417a3ab7dfd3c5f2609a10b6203"} Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.227218 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.262618 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.262709 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:08 crc kubenswrapper[4719]: E1215 12:34:08.262752 4719 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 15 12:34:08 crc kubenswrapper[4719]: E1215 12:34:08.262821 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:09.262801549 +0000 UTC m=+1010.205094579 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "metrics-server-cert" not found Dec 15 12:34:08 crc kubenswrapper[4719]: E1215 12:34:08.262917 4719 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 15 12:34:08 crc kubenswrapper[4719]: E1215 12:34:08.262988 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:09.262970274 +0000 UTC m=+1010.205263304 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "webhook-server-cert" not found Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.363677 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:08 crc kubenswrapper[4719]: E1215 12:34:08.364049 4719 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:08 crc kubenswrapper[4719]: E1215 12:34:08.364132 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert podName:b6efe10d-9999-4340-822b-48fb8fb99f14 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:10.364113892 +0000 UTC m=+1011.306406922 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert") pod "infra-operator-controller-manager-58944d7758-wh4dh" (UID: "b6efe10d-9999-4340-822b-48fb8fb99f14") : secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.537929 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6"] Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.556277 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-95949466-k2fz4"] Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.576996 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s"] Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.591947 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh"] Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.602125 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5"] Dec 15 12:34:08 crc kubenswrapper[4719]: W1215 12:34:08.626557 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0a679bd_fdb2_4a1f_97ae_8a31945f4c2e.slice/crio-cadab844d8ceae8ea290772cb396ed26dc09ca167d99bd51b7357891a3089cf5 WatchSource:0}: Error finding container cadab844d8ceae8ea290772cb396ed26dc09ca167d99bd51b7357891a3089cf5: Status 404 returned error can't find the container with id cadab844d8ceae8ea290772cb396ed26dc09ca167d99bd51b7357891a3089cf5 Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.633493 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9"] Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.647545 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f"] Dec 15 12:34:08 crc kubenswrapper[4719]: E1215 12:34:08.666718 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qm699,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-7cd87b778f-hb66s_openstack-operators(83a0cff1-8c43-4bb1-aa4f-66c4662fd235): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.667397 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj"] Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.669330 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:08 crc kubenswrapper[4719]: E1215 12:34:08.669480 4719 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:08 crc kubenswrapper[4719]: E1215 12:34:08.669537 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert podName:24c5b173-c013-45e1-9563-1e49556b23a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:10.669506814 +0000 UTC m=+1011.611799834 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" (UID: "24c5b173-c013-45e1-9563-1e49556b23a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:08 crc kubenswrapper[4719]: E1215 12:34:08.670590 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" podUID="83a0cff1-8c43-4bb1-aa4f-66c4662fd235" Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.675464 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg"] Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.684840 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s"] Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.856497 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl"] Dec 15 12:34:08 crc kubenswrapper[4719]: I1215 12:34:08.926761 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67"] Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.007182 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-767f9d7567-2866h"] Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.015615 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4"] Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.035312 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t"] Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.085752 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vmcmd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-756ccf86c7-l98v4_openstack-operators(72cd4bb7-5783-492b-bae4-1a9fa633cd7b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.086822 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" podUID="72cd4bb7-5783-492b-bae4-1a9fa633cd7b" Dec 15 12:34:09 crc kubenswrapper[4719]: W1215 12:34:09.090112 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod345b77d5_0419_433e_b7d4_053a0b6fcf86.slice/crio-888f2b195cb8ef9dbdf0a43560120e7e7cf1424b5cb5bef782e1fc8a5db7eaba WatchSource:0}: Error finding container 888f2b195cb8ef9dbdf0a43560120e7e7cf1424b5cb5bef782e1fc8a5db7eaba: Status 404 returned error can't find the container with id 888f2b195cb8ef9dbdf0a43560120e7e7cf1424b5cb5bef782e1fc8a5db7eaba Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.098375 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wzwgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-jzg2t_openstack-operators(0c36396e-131f-4154-b82e-490b36d7fa63): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.099670 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" podUID="0c36396e-131f-4154-b82e-490b36d7fa63" Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.099927 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vq2n9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-767f9d7567-2866h_openstack-operators(345b77d5-0419-433e-b7d4-053a0b6fcf86): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.101133 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" podUID="345b77d5-0419-433e-b7d4-053a0b6fcf86" Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.132655 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp"] Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.147498 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" event={"ID":"571dd62a-58e8-4dab-ad04-a95621a65078","Type":"ContainerStarted","Data":"4baca6d8f4b45b79cf52a70945d41094e6ec5d79d5eeab6709fa603f5611c5ba"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.148756 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" event={"ID":"9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc","Type":"ContainerStarted","Data":"264af9c69f6bc24bcb1c924970a6fd3b4eea88ce29e9ea9511921a0f7d47f24a"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.149343 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" event={"ID":"c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157","Type":"ContainerStarted","Data":"e569f80a3c808c08acf8d27f618bc67b2d99d3bb24d30a9423b807a6d86917da"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.150133 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" event={"ID":"375826ba-0785-4935-96c2-a2e5345aaf45","Type":"ContainerStarted","Data":"e99e6cb6e88f51b6692af467f830c4d11231c04b3da875ffee88780960c3f4e2"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.150707 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" event={"ID":"0c36396e-131f-4154-b82e-490b36d7fa63","Type":"ContainerStarted","Data":"3ca7a1c20889fa53c12831369d5a80ad381ce4a950a8959da5132bd638e06453"} Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.152069 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" podUID="0c36396e-131f-4154-b82e-490b36d7fa63" Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.185788 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" event={"ID":"345b77d5-0419-433e-b7d4-053a0b6fcf86","Type":"ContainerStarted","Data":"888f2b195cb8ef9dbdf0a43560120e7e7cf1424b5cb5bef782e1fc8a5db7eaba"} Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.198670 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027\\\"\"" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" podUID="345b77d5-0419-433e-b7d4-053a0b6fcf86" Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.204247 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" event={"ID":"72cd4bb7-5783-492b-bae4-1a9fa633cd7b","Type":"ContainerStarted","Data":"92fda48aa7e99ff29c5d6331f914fbf9b21af1414ccee51e2704a1ad2abde5b1"} Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.206360 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" podUID="72cd4bb7-5783-492b-bae4-1a9fa633cd7b" Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.209492 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-95949466-k2fz4" event={"ID":"5c7fa683-b59c-45ef-bc29-92400f43e4b8","Type":"ContainerStarted","Data":"50b46d29c05e70ffb2a1ae40204e04ef6058c98307abc419b4e8150c41bb12bf"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.214871 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" event={"ID":"83a0cff1-8c43-4bb1-aa4f-66c4662fd235","Type":"ContainerStarted","Data":"3cbfe3fd8125b3680398af73d569008e7cf3742bcdd40f6c9ca16870b063a4a3"} Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.216273 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" podUID="83a0cff1-8c43-4bb1-aa4f-66c4662fd235" Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.223932 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f" event={"ID":"3b66851a-75d1-46ea-8d06-ad7385fdff6b","Type":"ContainerStarted","Data":"428c5d625667c437936b18a0fa3960b81f4d3921f7c4dcead83c69b601653603"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.242149 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" event={"ID":"9adcd81c-e72b-44b8-bd4f-36e49c8e5c82","Type":"ContainerStarted","Data":"64625494b58a8b6c3f676bdde495e07a66706c59813627bd90e10aebf5f7f096"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.251077 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" event={"ID":"def1648e-ccf3-4a95-9a44-92bf0f83cb49","Type":"ContainerStarted","Data":"efa638d7532dc567fe4c4f15394cfcc822be0d014a21c90540443c822fdbb16a"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.264217 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" event={"ID":"e6adc1e0-5c73-431d-a94d-e2eabc16dd03","Type":"ContainerStarted","Data":"141cd7845a30b90bebcc54caa5932aff3d7866098cc420a2e64c6b9ba4b453da"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.275225 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" event={"ID":"828d7a4d-0956-40fc-967f-e52d1701953c","Type":"ContainerStarted","Data":"e915f3237e813ec6fe50cdc5625156d38260d29cf4f64826aef66368c034f03f"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.282086 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" event={"ID":"d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e","Type":"ContainerStarted","Data":"cadab844d8ceae8ea290772cb396ed26dc09ca167d99bd51b7357891a3089cf5"} Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.300138 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:09 crc kubenswrapper[4719]: I1215 12:34:09.300445 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.303360 4719 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.303396 4719 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.303424 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:11.303405723 +0000 UTC m=+1012.245698753 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "webhook-server-cert" not found Dec 15 12:34:09 crc kubenswrapper[4719]: E1215 12:34:09.303448 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:11.303432364 +0000 UTC m=+1012.245725394 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "metrics-server-cert" not found Dec 15 12:34:10 crc kubenswrapper[4719]: I1215 12:34:10.331126 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" event={"ID":"a9ebdf36-2095-47b8-8edf-89036ffdee84","Type":"ContainerStarted","Data":"088c0c54fe6140636183c0ebf77225125ae16f9fa7351a7e60bc7e42e222f893"} Dec 15 12:34:10 crc kubenswrapper[4719]: E1215 12:34:10.338702 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" podUID="83a0cff1-8c43-4bb1-aa4f-66c4662fd235" Dec 15 12:34:10 crc kubenswrapper[4719]: E1215 12:34:10.338789 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" podUID="72cd4bb7-5783-492b-bae4-1a9fa633cd7b" Dec 15 12:34:10 crc kubenswrapper[4719]: E1215 12:34:10.338845 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" podUID="0c36396e-131f-4154-b82e-490b36d7fa63" Dec 15 12:34:10 crc kubenswrapper[4719]: E1215 12:34:10.338900 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027\\\"\"" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" podUID="345b77d5-0419-433e-b7d4-053a0b6fcf86" Dec 15 12:34:10 crc kubenswrapper[4719]: I1215 12:34:10.453146 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:10 crc kubenswrapper[4719]: E1215 12:34:10.453310 4719 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:10 crc kubenswrapper[4719]: E1215 12:34:10.453358 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert podName:b6efe10d-9999-4340-822b-48fb8fb99f14 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:14.453338788 +0000 UTC m=+1015.395631818 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert") pod "infra-operator-controller-manager-58944d7758-wh4dh" (UID: "b6efe10d-9999-4340-822b-48fb8fb99f14") : secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:10 crc kubenswrapper[4719]: I1215 12:34:10.762992 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:10 crc kubenswrapper[4719]: E1215 12:34:10.763299 4719 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:10 crc kubenswrapper[4719]: E1215 12:34:10.763429 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert podName:24c5b173-c013-45e1-9563-1e49556b23a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:14.763376994 +0000 UTC m=+1015.705670024 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" (UID: "24c5b173-c013-45e1-9563-1e49556b23a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:11 crc kubenswrapper[4719]: I1215 12:34:11.373996 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:11 crc kubenswrapper[4719]: I1215 12:34:11.374182 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:11 crc kubenswrapper[4719]: E1215 12:34:11.374350 4719 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 15 12:34:11 crc kubenswrapper[4719]: E1215 12:34:11.374397 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:15.374382665 +0000 UTC m=+1016.316675695 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "metrics-server-cert" not found Dec 15 12:34:11 crc kubenswrapper[4719]: E1215 12:34:11.374739 4719 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 15 12:34:11 crc kubenswrapper[4719]: E1215 12:34:11.374775 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:15.374765257 +0000 UTC m=+1016.317058287 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "webhook-server-cert" not found Dec 15 12:34:14 crc kubenswrapper[4719]: I1215 12:34:14.533998 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:14 crc kubenswrapper[4719]: E1215 12:34:14.534214 4719 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:14 crc kubenswrapper[4719]: E1215 12:34:14.534918 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert podName:b6efe10d-9999-4340-822b-48fb8fb99f14 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:22.534897384 +0000 UTC m=+1023.477190404 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert") pod "infra-operator-controller-manager-58944d7758-wh4dh" (UID: "b6efe10d-9999-4340-822b-48fb8fb99f14") : secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:14 crc kubenswrapper[4719]: I1215 12:34:14.837094 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:14 crc kubenswrapper[4719]: E1215 12:34:14.837273 4719 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:14 crc kubenswrapper[4719]: E1215 12:34:14.837362 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert podName:24c5b173-c013-45e1-9563-1e49556b23a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:22.837343565 +0000 UTC m=+1023.779636595 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" (UID: "24c5b173-c013-45e1-9563-1e49556b23a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:15 crc kubenswrapper[4719]: I1215 12:34:15.444828 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:15 crc kubenswrapper[4719]: I1215 12:34:15.444961 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:15 crc kubenswrapper[4719]: E1215 12:34:15.445127 4719 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 15 12:34:15 crc kubenswrapper[4719]: E1215 12:34:15.445172 4719 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 15 12:34:15 crc kubenswrapper[4719]: E1215 12:34:15.445215 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:23.445192249 +0000 UTC m=+1024.387485289 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "metrics-server-cert" not found Dec 15 12:34:15 crc kubenswrapper[4719]: E1215 12:34:15.445251 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:23.44522808 +0000 UTC m=+1024.387521180 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "webhook-server-cert" not found Dec 15 12:34:22 crc kubenswrapper[4719]: I1215 12:34:22.547501 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:22 crc kubenswrapper[4719]: E1215 12:34:22.547811 4719 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:22 crc kubenswrapper[4719]: E1215 12:34:22.548481 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert podName:b6efe10d-9999-4340-822b-48fb8fb99f14 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:38.548451103 +0000 UTC m=+1039.490744123 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert") pod "infra-operator-controller-manager-58944d7758-wh4dh" (UID: "b6efe10d-9999-4340-822b-48fb8fb99f14") : secret "infra-operator-webhook-server-cert" not found Dec 15 12:34:22 crc kubenswrapper[4719]: I1215 12:34:22.853750 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:22 crc kubenswrapper[4719]: E1215 12:34:22.853957 4719 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:22 crc kubenswrapper[4719]: E1215 12:34:22.854042 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert podName:24c5b173-c013-45e1-9563-1e49556b23a5 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:38.854023321 +0000 UTC m=+1039.796316351 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" (UID: "24c5b173-c013-45e1-9563-1e49556b23a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 15 12:34:23 crc kubenswrapper[4719]: I1215 12:34:23.462961 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:23 crc kubenswrapper[4719]: I1215 12:34:23.463285 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:23 crc kubenswrapper[4719]: E1215 12:34:23.463116 4719 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 15 12:34:23 crc kubenswrapper[4719]: E1215 12:34:23.463438 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:39.463425934 +0000 UTC m=+1040.405718964 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "webhook-server-cert" not found Dec 15 12:34:23 crc kubenswrapper[4719]: E1215 12:34:23.463398 4719 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 15 12:34:23 crc kubenswrapper[4719]: E1215 12:34:23.463466 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs podName:db19bab3-733c-4ef1-8f46-df4cf5842701 nodeName:}" failed. No retries permitted until 2025-12-15 12:34:39.463461495 +0000 UTC m=+1040.405754525 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs") pod "openstack-operator-controller-manager-7b4b554f64-x55hj" (UID: "db19bab3-733c-4ef1-8f46-df4cf5842701") : secret "metrics-server-cert" not found Dec 15 12:34:24 crc kubenswrapper[4719]: E1215 12:34:24.642634 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad" Dec 15 12:34:24 crc kubenswrapper[4719]: E1215 12:34:24.643057 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7hjpg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-f76f4954c-zhvw6_openstack-operators(def1648e-ccf3-4a95-9a44-92bf0f83cb49): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:24 crc kubenswrapper[4719]: E1215 12:34:24.644196 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" podUID="def1648e-ccf3-4a95-9a44-92bf0f83cb49" Dec 15 12:34:25 crc kubenswrapper[4719]: E1215 12:34:25.459500 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" podUID="def1648e-ccf3-4a95-9a44-92bf0f83cb49" Dec 15 12:34:26 crc kubenswrapper[4719]: E1215 12:34:26.917891 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 15 12:34:26 crc kubenswrapper[4719]: E1215 12:34:26.918430 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qc9q8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-6ccf486b9-scntm_openstack-operators(15643ba0-8573-4f51-8ee6-9fd78e10b6a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:26 crc kubenswrapper[4719]: E1215 12:34:26.919687 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" podUID="15643ba0-8573-4f51-8ee6-9fd78e10b6a0" Dec 15 12:34:27 crc kubenswrapper[4719]: E1215 12:34:27.012453 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 15 12:34:27 crc kubenswrapper[4719]: E1215 12:34:27.012758 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vk7gs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-bf6d4f946-gn5n9_openstack-operators(d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:27 crc kubenswrapper[4719]: E1215 12:34:27.014195 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" podUID="d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e" Dec 15 12:34:27 crc kubenswrapper[4719]: E1215 12:34:27.473143 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" podUID="d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e" Dec 15 12:34:27 crc kubenswrapper[4719]: E1215 12:34:27.473637 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" podUID="15643ba0-8573-4f51-8ee6-9fd78e10b6a0" Dec 15 12:34:28 crc kubenswrapper[4719]: E1215 12:34:28.907039 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 15 12:34:28 crc kubenswrapper[4719]: E1215 12:34:28.907267 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bw2w6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8665b56d78-wslbg_openstack-operators(9adcd81c-e72b-44b8-bd4f-36e49c8e5c82): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:28 crc kubenswrapper[4719]: E1215 12:34:28.908400 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" podUID="9adcd81c-e72b-44b8-bd4f-36e49c8e5c82" Dec 15 12:34:29 crc kubenswrapper[4719]: E1215 12:34:29.486180 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" podUID="9adcd81c-e72b-44b8-bd4f-36e49c8e5c82" Dec 15 12:34:30 crc kubenswrapper[4719]: E1215 12:34:30.627447 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 15 12:34:30 crc kubenswrapper[4719]: E1215 12:34:30.627618 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mnrcb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5c6df8f9-d9d67_openstack-operators(c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:30 crc kubenswrapper[4719]: E1215 12:34:30.629298 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" podUID="c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157" Dec 15 12:34:31 crc kubenswrapper[4719]: E1215 12:34:31.293290 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a" Dec 15 12:34:31 crc kubenswrapper[4719]: E1215 12:34:31.293481 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pwncv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5fdd9786f7-89zrh_openstack-operators(571dd62a-58e8-4dab-ad04-a95621a65078): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:31 crc kubenswrapper[4719]: E1215 12:34:31.294660 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" podUID="571dd62a-58e8-4dab-ad04-a95621a65078" Dec 15 12:34:31 crc kubenswrapper[4719]: E1215 12:34:31.501632 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" podUID="c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157" Dec 15 12:34:31 crc kubenswrapper[4719]: E1215 12:34:31.502722 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" podUID="571dd62a-58e8-4dab-ad04-a95621a65078" Dec 15 12:34:37 crc kubenswrapper[4719]: E1215 12:34:37.675094 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f" Dec 15 12:34:37 crc kubenswrapper[4719]: E1215 12:34:37.676381 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kfq4x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-97d456b9-qxhfl_openstack-operators(375826ba-0785-4935-96c2-a2e5345aaf45): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:37 crc kubenswrapper[4719]: E1215 12:34:37.678065 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" podUID="375826ba-0785-4935-96c2-a2e5345aaf45" Dec 15 12:34:38 crc kubenswrapper[4719]: E1215 12:34:38.107144 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 15 12:34:38 crc kubenswrapper[4719]: E1215 12:34:38.107564 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pgmf5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-59b8dcb766-llbls_openstack-operators(6cdd3328-cd4d-4e90-8ab9-d0953cab73de): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:38 crc kubenswrapper[4719]: E1215 12:34:38.108747 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" podUID="6cdd3328-cd4d-4e90-8ab9-d0953cab73de" Dec 15 12:34:38 crc kubenswrapper[4719]: E1215 12:34:38.544050 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429\\\"\"" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" podUID="6cdd3328-cd4d-4e90-8ab9-d0953cab73de" Dec 15 12:34:38 crc kubenswrapper[4719]: E1215 12:34:38.544369 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" podUID="375826ba-0785-4935-96c2-a2e5345aaf45" Dec 15 12:34:38 crc kubenswrapper[4719]: I1215 12:34:38.589042 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:38 crc kubenswrapper[4719]: I1215 12:34:38.594512 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6efe10d-9999-4340-822b-48fb8fb99f14-cert\") pod \"infra-operator-controller-manager-58944d7758-wh4dh\" (UID: \"b6efe10d-9999-4340-822b-48fb8fb99f14\") " pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:38 crc kubenswrapper[4719]: I1215 12:34:38.756036 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-h4nt9" Dec 15 12:34:38 crc kubenswrapper[4719]: I1215 12:34:38.764580 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:38 crc kubenswrapper[4719]: I1215 12:34:38.894769 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:38 crc kubenswrapper[4719]: I1215 12:34:38.899147 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24c5b173-c013-45e1-9563-1e49556b23a5-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556\" (UID: \"24c5b173-c013-45e1-9563-1e49556b23a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:38 crc kubenswrapper[4719]: I1215 12:34:38.917335 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8f4t2" Dec 15 12:34:38 crc kubenswrapper[4719]: I1215 12:34:38.925203 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:39 crc kubenswrapper[4719]: I1215 12:34:39.485149 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 12:34:39 crc kubenswrapper[4719]: I1215 12:34:39.504317 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:39 crc kubenswrapper[4719]: I1215 12:34:39.504390 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:39 crc kubenswrapper[4719]: I1215 12:34:39.513996 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-webhook-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:39 crc kubenswrapper[4719]: I1215 12:34:39.520750 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db19bab3-733c-4ef1-8f46-df4cf5842701-metrics-certs\") pod \"openstack-operator-controller-manager-7b4b554f64-x55hj\" (UID: \"db19bab3-733c-4ef1-8f46-df4cf5842701\") " pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:39 crc kubenswrapper[4719]: I1215 12:34:39.729179 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-kfj4g" Dec 15 12:34:39 crc kubenswrapper[4719]: I1215 12:34:39.737934 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:39 crc kubenswrapper[4719]: E1215 12:34:39.978430 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a" Dec 15 12:34:39 crc kubenswrapper[4719]: E1215 12:34:39.978682 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qmgrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-55f78b7c4c-292vp_openstack-operators(a9ebdf36-2095-47b8-8edf-89036ffdee84): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:39 crc kubenswrapper[4719]: E1215 12:34:39.981350 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" podUID="a9ebdf36-2095-47b8-8edf-89036ffdee84" Dec 15 12:34:40 crc kubenswrapper[4719]: E1215 12:34:40.564031 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" podUID="a9ebdf36-2095-47b8-8edf-89036ffdee84" Dec 15 12:34:41 crc kubenswrapper[4719]: E1215 12:34:41.994355 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 15 12:34:41 crc kubenswrapper[4719]: E1215 12:34:41.995906 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-btwv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-5c7cbf548f-5slz5_openstack-operators(9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:41 crc kubenswrapper[4719]: E1215 12:34:41.997841 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" podUID="9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc" Dec 15 12:34:42 crc kubenswrapper[4719]: E1215 12:34:42.415061 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 15 12:34:42 crc kubenswrapper[4719]: E1215 12:34:42.415596 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vmcmd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-756ccf86c7-l98v4_openstack-operators(72cd4bb7-5783-492b-bae4-1a9fa633cd7b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:42 crc kubenswrapper[4719]: E1215 12:34:42.416766 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" podUID="72cd4bb7-5783-492b-bae4-1a9fa633cd7b" Dec 15 12:34:42 crc kubenswrapper[4719]: E1215 12:34:42.576333 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" podUID="9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc" Dec 15 12:34:42 crc kubenswrapper[4719]: E1215 12:34:42.913880 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 15 12:34:42 crc kubenswrapper[4719]: E1215 12:34:42.914277 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qm699,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-7cd87b778f-hb66s_openstack-operators(83a0cff1-8c43-4bb1-aa4f-66c4662fd235): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:42 crc kubenswrapper[4719]: E1215 12:34:42.915487 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" podUID="83a0cff1-8c43-4bb1-aa4f-66c4662fd235" Dec 15 12:34:43 crc kubenswrapper[4719]: E1215 12:34:43.387323 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3" Dec 15 12:34:43 crc kubenswrapper[4719]: E1215 12:34:43.388128 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n65v2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-5cf45c46bd-pmmwk_openstack-operators(7f47c007-e5bf-47dd-8905-69ed7cfc3a9a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:43 crc kubenswrapper[4719]: E1215 12:34:43.390005 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" podUID="7f47c007-e5bf-47dd-8905-69ed7cfc3a9a" Dec 15 12:34:43 crc kubenswrapper[4719]: E1215 12:34:43.583145 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" podUID="7f47c007-e5bf-47dd-8905-69ed7cfc3a9a" Dec 15 12:34:43 crc kubenswrapper[4719]: E1215 12:34:43.754083 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 15 12:34:43 crc kubenswrapper[4719]: E1215 12:34:43.754298 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wzwgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-jzg2t_openstack-operators(0c36396e-131f-4154-b82e-490b36d7fa63): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:43 crc kubenswrapper[4719]: E1215 12:34:43.755469 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" podUID="0c36396e-131f-4154-b82e-490b36d7fa63" Dec 15 12:34:44 crc kubenswrapper[4719]: E1215 12:34:44.224689 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 15 12:34:44 crc kubenswrapper[4719]: E1215 12:34:44.224883 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ghmnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5fbbf8b6cc-c9szj_openstack-operators(828d7a4d-0956-40fc-967f-e52d1701953c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:34:44 crc kubenswrapper[4719]: E1215 12:34:44.227111 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" podUID="828d7a4d-0956-40fc-967f-e52d1701953c" Dec 15 12:34:44 crc kubenswrapper[4719]: E1215 12:34:44.612709 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" podUID="828d7a4d-0956-40fc-967f-e52d1701953c" Dec 15 12:34:44 crc kubenswrapper[4719]: I1215 12:34:44.743331 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556"] Dec 15 12:34:44 crc kubenswrapper[4719]: I1215 12:34:44.792639 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj"] Dec 15 12:34:44 crc kubenswrapper[4719]: I1215 12:34:44.882257 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh"] Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.597514 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" event={"ID":"db19bab3-733c-4ef1-8f46-df4cf5842701","Type":"ContainerStarted","Data":"65c146a371e04b5c548c0519fb6d6d95170f8998d83260436fb4f96ff2fe5f19"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.597891 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" event={"ID":"db19bab3-733c-4ef1-8f46-df4cf5842701","Type":"ContainerStarted","Data":"49b169b4028409d4c23baccbe8355c17f32bebf6d10472b3444009032cb8381b"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.598837 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.606317 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f" event={"ID":"3b66851a-75d1-46ea-8d06-ad7385fdff6b","Type":"ContainerStarted","Data":"cbffc18e32fd05a41ef8ce4333a47a3bfedad3f39e6c5e7cb1223de7e1c32f85"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.606895 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.608096 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v" event={"ID":"7ced32f3-ca3c-406c-966b-93be87cd6d25","Type":"ContainerStarted","Data":"5b40b105e8289a305c9e63ffc81d85e9a8be6c0172d1f14a0b3a859703dac6cd"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.608519 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.609655 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" event={"ID":"e6adc1e0-5c73-431d-a94d-e2eabc16dd03","Type":"ContainerStarted","Data":"69aa912ef96b81760133e8532830f9ed726f6ae96b0d412060b7bc4f34665be3"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.610232 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.615080 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" event={"ID":"b6efe10d-9999-4340-822b-48fb8fb99f14","Type":"ContainerStarted","Data":"72dcc3fe56da7f442061048fab74ff4bebde47d8cbbce113bb41906d0e27221e"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.618447 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" event={"ID":"d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e","Type":"ContainerStarted","Data":"4d3844a96997cf6db1ffa2279a48f03db87758df68f7e7bba06b7c8073c4714e"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.619481 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.624039 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" event={"ID":"24c5b173-c013-45e1-9563-1e49556b23a5","Type":"ContainerStarted","Data":"3b43e2974245cd9979e368959e599cb0a5e3cd1b281cfbb334a11fda292121b9"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.629124 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" event={"ID":"345b77d5-0419-433e-b7d4-053a0b6fcf86","Type":"ContainerStarted","Data":"625c34e479593572c61dce677831b9f7b064480fd181a55e4c83405a5bdc4eb2"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.630252 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.636280 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" podStartSLOduration=38.636264138 podStartE2EDuration="38.636264138s" podCreationTimestamp="2025-12-15 12:34:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:34:45.62894455 +0000 UTC m=+1046.571237600" watchObservedRunningTime="2025-12-15 12:34:45.636264138 +0000 UTC m=+1046.578557168" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.637595 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" event={"ID":"9adcd81c-e72b-44b8-bd4f-36e49c8e5c82","Type":"ContainerStarted","Data":"2e06f5c444dea87b15d712841b387f8f50219164a94447abb1656c43fdd9d703"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.638335 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.640994 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" event={"ID":"def1648e-ccf3-4a95-9a44-92bf0f83cb49","Type":"ContainerStarted","Data":"6ac999164babd2523fbebc8e097e4c246c8ee09bb670b3eb14fb1d6b8bfabe8a"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.641595 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.644596 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-95949466-k2fz4" event={"ID":"5c7fa683-b59c-45ef-bc29-92400f43e4b8","Type":"ContainerStarted","Data":"2258e41bb504fb46d3b1627c9cd695fffc7da9b2f0aaa76ca8b8d861eeb21de4"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.645354 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-95949466-k2fz4" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.649823 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" event={"ID":"15643ba0-8573-4f51-8ee6-9fd78e10b6a0","Type":"ContainerStarted","Data":"1385913be4eab52d2a817dd3e0d29ca43400ced29e74f2acd6c3d6cff1c66b94"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.650547 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.653619 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" event={"ID":"c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157","Type":"ContainerStarted","Data":"97a320b96716c2a39c27e209d3fc739b6fd0e9fb006ac2a0e3ef6fd15ed3e9a9"} Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.654250 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.716252 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" podStartSLOduration=4.551246251 podStartE2EDuration="39.716239349s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.568551803 +0000 UTC m=+1009.510844833" lastFinishedPulling="2025-12-15 12:34:43.733544901 +0000 UTC m=+1044.675837931" observedRunningTime="2025-12-15 12:34:45.678128312 +0000 UTC m=+1046.620421342" watchObservedRunningTime="2025-12-15 12:34:45.716239349 +0000 UTC m=+1046.658532379" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.842232 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" podStartSLOduration=4.047850102 podStartE2EDuration="39.842211901s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.630331143 +0000 UTC m=+1009.572624173" lastFinishedPulling="2025-12-15 12:34:44.424692942 +0000 UTC m=+1045.366985972" observedRunningTime="2025-12-15 12:34:45.795682573 +0000 UTC m=+1046.737975603" watchObservedRunningTime="2025-12-15 12:34:45.842211901 +0000 UTC m=+1046.784504931" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.844739 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v" podStartSLOduration=4.052307941 podStartE2EDuration="39.844724139s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:07.941186455 +0000 UTC m=+1008.883479485" lastFinishedPulling="2025-12-15 12:34:43.733602653 +0000 UTC m=+1044.675895683" observedRunningTime="2025-12-15 12:34:45.838502935 +0000 UTC m=+1046.780795975" watchObservedRunningTime="2025-12-15 12:34:45.844724139 +0000 UTC m=+1046.787017169" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.897280 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f" podStartSLOduration=4.808094701 podStartE2EDuration="39.897263005s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.64445741 +0000 UTC m=+1009.586750440" lastFinishedPulling="2025-12-15 12:34:43.733625714 +0000 UTC m=+1044.675918744" observedRunningTime="2025-12-15 12:34:45.895538602 +0000 UTC m=+1046.837831652" watchObservedRunningTime="2025-12-15 12:34:45.897263005 +0000 UTC m=+1046.839556035" Dec 15 12:34:45 crc kubenswrapper[4719]: I1215 12:34:45.964111 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-95949466-k2fz4" podStartSLOduration=4.799318225 podStartE2EDuration="39.964093226s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.568808341 +0000 UTC m=+1009.511101381" lastFinishedPulling="2025-12-15 12:34:43.733583352 +0000 UTC m=+1044.675876382" observedRunningTime="2025-12-15 12:34:45.958964986 +0000 UTC m=+1046.901258026" watchObservedRunningTime="2025-12-15 12:34:45.964093226 +0000 UTC m=+1046.906386256" Dec 15 12:34:46 crc kubenswrapper[4719]: I1215 12:34:46.076281 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" podStartSLOduration=4.233777854 podStartE2EDuration="40.076262869s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.574718354 +0000 UTC m=+1009.517011384" lastFinishedPulling="2025-12-15 12:34:44.417203369 +0000 UTC m=+1045.359496399" observedRunningTime="2025-12-15 12:34:46.071199721 +0000 UTC m=+1047.013492751" watchObservedRunningTime="2025-12-15 12:34:46.076262869 +0000 UTC m=+1047.018555899" Dec 15 12:34:46 crc kubenswrapper[4719]: I1215 12:34:46.079317 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" podStartSLOduration=4.944882971 podStartE2EDuration="40.079306044s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:09.099742776 +0000 UTC m=+1010.042035806" lastFinishedPulling="2025-12-15 12:34:44.234165849 +0000 UTC m=+1045.176458879" observedRunningTime="2025-12-15 12:34:46.015413824 +0000 UTC m=+1046.957706854" watchObservedRunningTime="2025-12-15 12:34:46.079306044 +0000 UTC m=+1047.021599074" Dec 15 12:34:46 crc kubenswrapper[4719]: I1215 12:34:46.165214 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" podStartSLOduration=4.30287785 podStartE2EDuration="40.165198649s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.649155225 +0000 UTC m=+1009.591448255" lastFinishedPulling="2025-12-15 12:34:44.511476014 +0000 UTC m=+1045.453769054" observedRunningTime="2025-12-15 12:34:46.114388596 +0000 UTC m=+1047.056681636" watchObservedRunningTime="2025-12-15 12:34:46.165198649 +0000 UTC m=+1047.107491679" Dec 15 12:34:46 crc kubenswrapper[4719]: I1215 12:34:46.166127 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" podStartSLOduration=3.855786116 podStartE2EDuration="40.166124007s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.106041882 +0000 UTC m=+1009.048334912" lastFinishedPulling="2025-12-15 12:34:44.416379773 +0000 UTC m=+1045.358672803" observedRunningTime="2025-12-15 12:34:46.160615465 +0000 UTC m=+1047.102908495" watchObservedRunningTime="2025-12-15 12:34:46.166124007 +0000 UTC m=+1047.108417037" Dec 15 12:34:46 crc kubenswrapper[4719]: I1215 12:34:46.506119 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" podStartSLOduration=4.935723529 podStartE2EDuration="40.506104214s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.950317617 +0000 UTC m=+1009.892610647" lastFinishedPulling="2025-12-15 12:34:44.520698302 +0000 UTC m=+1045.462991332" observedRunningTime="2025-12-15 12:34:46.199374582 +0000 UTC m=+1047.141667602" watchObservedRunningTime="2025-12-15 12:34:46.506104214 +0000 UTC m=+1047.448397244" Dec 15 12:34:49 crc kubenswrapper[4719]: I1215 12:34:49.677604 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" event={"ID":"571dd62a-58e8-4dab-ad04-a95621a65078","Type":"ContainerStarted","Data":"f5f5638da58aee1ae27ea994b36e4a2dac0b571ce79199d42b677fd11ce60e36"} Dec 15 12:34:49 crc kubenswrapper[4719]: I1215 12:34:49.678536 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" Dec 15 12:34:49 crc kubenswrapper[4719]: I1215 12:34:49.696939 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" podStartSLOduration=3.310053064 podStartE2EDuration="43.696850187s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.557012776 +0000 UTC m=+1009.499305806" lastFinishedPulling="2025-12-15 12:34:48.943809889 +0000 UTC m=+1049.886102929" observedRunningTime="2025-12-15 12:34:49.695082822 +0000 UTC m=+1050.637375852" watchObservedRunningTime="2025-12-15 12:34:49.696850187 +0000 UTC m=+1050.639143217" Dec 15 12:34:50 crc kubenswrapper[4719]: I1215 12:34:50.687226 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" event={"ID":"24c5b173-c013-45e1-9563-1e49556b23a5","Type":"ContainerStarted","Data":"19a043ed06b5fb47e0ca0c07fe4799d2393f2d77ffd5fb728abeb7b0465c92dc"} Dec 15 12:34:50 crc kubenswrapper[4719]: I1215 12:34:50.724170 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" podStartSLOduration=39.783230393 podStartE2EDuration="44.724147855s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:44.816031097 +0000 UTC m=+1045.758324127" lastFinishedPulling="2025-12-15 12:34:49.756948549 +0000 UTC m=+1050.699241589" observedRunningTime="2025-12-15 12:34:50.718490809 +0000 UTC m=+1051.660783839" watchObservedRunningTime="2025-12-15 12:34:50.724147855 +0000 UTC m=+1051.666440895" Dec 15 12:34:51 crc kubenswrapper[4719]: I1215 12:34:51.694048 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:52 crc kubenswrapper[4719]: I1215 12:34:52.701540 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" event={"ID":"375826ba-0785-4935-96c2-a2e5345aaf45","Type":"ContainerStarted","Data":"883a46c8104afd0cb955527291448501387b36c2c392b2d26ab57487188d3644"} Dec 15 12:34:52 crc kubenswrapper[4719]: I1215 12:34:52.703935 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" Dec 15 12:34:52 crc kubenswrapper[4719]: I1215 12:34:52.705846 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" event={"ID":"b6efe10d-9999-4340-822b-48fb8fb99f14","Type":"ContainerStarted","Data":"0a0e786736b3023595a92fbe6648be38296939f9371b2c51fff6f2aa4855dcf2"} Dec 15 12:34:52 crc kubenswrapper[4719]: I1215 12:34:52.706014 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:52 crc kubenswrapper[4719]: I1215 12:34:52.707416 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" event={"ID":"6cdd3328-cd4d-4e90-8ab9-d0953cab73de","Type":"ContainerStarted","Data":"40236e90d1b9e5cab5e3f4a46f2dab624179d0b8199ad15c949a5ad328d9e3a7"} Dec 15 12:34:52 crc kubenswrapper[4719]: I1215 12:34:52.729201 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" podStartSLOduration=3.61020176 podStartE2EDuration="46.729182518s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.876321708 +0000 UTC m=+1009.818614738" lastFinishedPulling="2025-12-15 12:34:51.995302466 +0000 UTC m=+1052.937595496" observedRunningTime="2025-12-15 12:34:52.720933411 +0000 UTC m=+1053.663226431" watchObservedRunningTime="2025-12-15 12:34:52.729182518 +0000 UTC m=+1053.671475558" Dec 15 12:34:52 crc kubenswrapper[4719]: I1215 12:34:52.744742 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" podStartSLOduration=4.686327766 podStartE2EDuration="46.744721962s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:07.936826041 +0000 UTC m=+1008.879119071" lastFinishedPulling="2025-12-15 12:34:49.995220237 +0000 UTC m=+1050.937513267" observedRunningTime="2025-12-15 12:34:52.743341599 +0000 UTC m=+1053.685634619" watchObservedRunningTime="2025-12-15 12:34:52.744721962 +0000 UTC m=+1053.687014992" Dec 15 12:34:52 crc kubenswrapper[4719]: I1215 12:34:52.764246 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" podStartSLOduration=39.952730562 podStartE2EDuration="46.764229589s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:44.931001757 +0000 UTC m=+1045.873294787" lastFinishedPulling="2025-12-15 12:34:51.742500784 +0000 UTC m=+1052.684793814" observedRunningTime="2025-12-15 12:34:52.76167612 +0000 UTC m=+1053.703969170" watchObservedRunningTime="2025-12-15 12:34:52.764229589 +0000 UTC m=+1053.706522629" Dec 15 12:34:54 crc kubenswrapper[4719]: E1215 12:34:54.477882 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" podUID="83a0cff1-8c43-4bb1-aa4f-66c4662fd235" Dec 15 12:34:54 crc kubenswrapper[4719]: I1215 12:34:54.730031 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" event={"ID":"a9ebdf36-2095-47b8-8edf-89036ffdee84","Type":"ContainerStarted","Data":"793edf1e3d53cafc3da33441e0de9c547c67cedd2722718f94cfe209ba4d1685"} Dec 15 12:34:54 crc kubenswrapper[4719]: I1215 12:34:54.731237 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" Dec 15 12:34:54 crc kubenswrapper[4719]: I1215 12:34:54.749560 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" podStartSLOduration=3.169011242 podStartE2EDuration="47.749542018s" podCreationTimestamp="2025-12-15 12:34:07 +0000 UTC" firstStartedPulling="2025-12-15 12:34:09.185058364 +0000 UTC m=+1010.127351394" lastFinishedPulling="2025-12-15 12:34:53.76558914 +0000 UTC m=+1054.707882170" observedRunningTime="2025-12-15 12:34:54.746604247 +0000 UTC m=+1055.688897277" watchObservedRunningTime="2025-12-15 12:34:54.749542018 +0000 UTC m=+1055.691835048" Dec 15 12:34:55 crc kubenswrapper[4719]: E1215 12:34:55.477307 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" podUID="72cd4bb7-5783-492b-bae4-1a9fa633cd7b" Dec 15 12:34:55 crc kubenswrapper[4719]: I1215 12:34:55.737176 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" event={"ID":"9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc","Type":"ContainerStarted","Data":"a62c81833e774e98e95bc3f3080c72fc0353d4c6e8ae613083d686bb9a303c16"} Dec 15 12:34:55 crc kubenswrapper[4719]: I1215 12:34:55.737724 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" Dec 15 12:34:55 crc kubenswrapper[4719]: I1215 12:34:55.752180 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" podStartSLOduration=3.130571832 podStartE2EDuration="49.752156308s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.571261386 +0000 UTC m=+1009.513554416" lastFinishedPulling="2025-12-15 12:34:55.192845862 +0000 UTC m=+1056.135138892" observedRunningTime="2025-12-15 12:34:55.751821307 +0000 UTC m=+1056.694114337" watchObservedRunningTime="2025-12-15 12:34:55.752156308 +0000 UTC m=+1056.694449338" Dec 15 12:34:56 crc kubenswrapper[4719]: E1215 12:34:56.478253 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" podUID="0c36396e-131f-4154-b82e-490b36d7fa63" Dec 15 12:34:56 crc kubenswrapper[4719]: I1215 12:34:56.795901 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-5td8v" Dec 15 12:34:56 crc kubenswrapper[4719]: I1215 12:34:56.907286 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" Dec 15 12:34:56 crc kubenswrapper[4719]: I1215 12:34:56.909474 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-llbls" Dec 15 12:34:56 crc kubenswrapper[4719]: I1215 12:34:56.934126 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-scntm" Dec 15 12:34:57 crc kubenswrapper[4719]: I1215 12:34:57.005896 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" Dec 15 12:34:57 crc kubenswrapper[4719]: I1215 12:34:57.032126 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-89zrh" Dec 15 12:34:57 crc kubenswrapper[4719]: I1215 12:34:57.070609 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-95949466-k2fz4" Dec 15 12:34:57 crc kubenswrapper[4719]: I1215 12:34:57.265153 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-zhvw6" Dec 15 12:34:57 crc kubenswrapper[4719]: I1215 12:34:57.336155 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-p7z6f" Dec 15 12:34:57 crc kubenswrapper[4719]: I1215 12:34:57.484022 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-wslbg" Dec 15 12:34:57 crc kubenswrapper[4719]: I1215 12:34:57.492834 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-gn5n9" Dec 15 12:34:57 crc kubenswrapper[4719]: I1215 12:34:57.683296 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-d9d67" Dec 15 12:34:57 crc kubenswrapper[4719]: I1215 12:34:57.694147 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-qxhfl" Dec 15 12:34:57 crc kubenswrapper[4719]: I1215 12:34:57.923743 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-2866h" Dec 15 12:34:58 crc kubenswrapper[4719]: I1215 12:34:58.760725 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" event={"ID":"7f47c007-e5bf-47dd-8905-69ed7cfc3a9a","Type":"ContainerStarted","Data":"bc3b9cf2b6761347e3c2f82c610e02ff42df720602a110293c017da020dce19e"} Dec 15 12:34:58 crc kubenswrapper[4719]: I1215 12:34:58.761197 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" Dec 15 12:34:58 crc kubenswrapper[4719]: I1215 12:34:58.771236 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-58944d7758-wh4dh" Dec 15 12:34:58 crc kubenswrapper[4719]: I1215 12:34:58.782595 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" podStartSLOduration=2.754980976 podStartE2EDuration="52.782570681s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:07.989000234 +0000 UTC m=+1008.931293274" lastFinishedPulling="2025-12-15 12:34:58.016589949 +0000 UTC m=+1058.958882979" observedRunningTime="2025-12-15 12:34:58.779376131 +0000 UTC m=+1059.721669191" watchObservedRunningTime="2025-12-15 12:34:58.782570681 +0000 UTC m=+1059.724863711" Dec 15 12:34:58 crc kubenswrapper[4719]: I1215 12:34:58.933253 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556" Dec 15 12:34:59 crc kubenswrapper[4719]: I1215 12:34:59.743974 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7b4b554f64-x55hj" Dec 15 12:34:59 crc kubenswrapper[4719]: I1215 12:34:59.767428 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" event={"ID":"828d7a4d-0956-40fc-967f-e52d1701953c","Type":"ContainerStarted","Data":"be2c5cfde6a67573d41fd554ec6a40fb4550806c4bcda3996bf783c04eb1998f"} Dec 15 12:34:59 crc kubenswrapper[4719]: I1215 12:34:59.767670 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" Dec 15 12:34:59 crc kubenswrapper[4719]: I1215 12:34:59.828700 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" podStartSLOduration=3.532890897 podStartE2EDuration="53.828680844s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.630291202 +0000 UTC m=+1009.572584232" lastFinishedPulling="2025-12-15 12:34:58.926081149 +0000 UTC m=+1059.868374179" observedRunningTime="2025-12-15 12:34:59.823910746 +0000 UTC m=+1060.766203796" watchObservedRunningTime="2025-12-15 12:34:59.828680844 +0000 UTC m=+1060.770973874" Dec 15 12:35:06 crc kubenswrapper[4719]: I1215 12:35:06.764161 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5cf45c46bd-pmmwk" Dec 15 12:35:07 crc kubenswrapper[4719]: I1215 12:35:07.220685 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-5slz5" Dec 15 12:35:07 crc kubenswrapper[4719]: I1215 12:35:07.388698 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-c9szj" Dec 15 12:35:07 crc kubenswrapper[4719]: I1215 12:35:07.743549 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-292vp" Dec 15 12:35:07 crc kubenswrapper[4719]: I1215 12:35:07.877173 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" event={"ID":"83a0cff1-8c43-4bb1-aa4f-66c4662fd235","Type":"ContainerStarted","Data":"b109e9dd5afed2b6a4df7af0a54d1fd16ab5f2839dd1b41d3b1c6ad4bde53e4a"} Dec 15 12:35:07 crc kubenswrapper[4719]: I1215 12:35:07.877427 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" Dec 15 12:35:07 crc kubenswrapper[4719]: I1215 12:35:07.892559 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" podStartSLOduration=3.382301171 podStartE2EDuration="1m1.892545888s" podCreationTimestamp="2025-12-15 12:34:06 +0000 UTC" firstStartedPulling="2025-12-15 12:34:08.666582444 +0000 UTC m=+1009.608875474" lastFinishedPulling="2025-12-15 12:35:07.176827161 +0000 UTC m=+1068.119120191" observedRunningTime="2025-12-15 12:35:07.890329998 +0000 UTC m=+1068.832623028" watchObservedRunningTime="2025-12-15 12:35:07.892545888 +0000 UTC m=+1068.834838908" Dec 15 12:35:10 crc kubenswrapper[4719]: I1215 12:35:10.908694 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" event={"ID":"72cd4bb7-5783-492b-bae4-1a9fa633cd7b","Type":"ContainerStarted","Data":"9c68fbd8501c129d79185bb07d0b26094b8cc1dc48beb3174c184d22359f9109"} Dec 15 12:35:10 crc kubenswrapper[4719]: I1215 12:35:10.909488 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" Dec 15 12:35:10 crc kubenswrapper[4719]: I1215 12:35:10.926117 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" podStartSLOduration=2.73335733 podStartE2EDuration="1m3.926094757s" podCreationTimestamp="2025-12-15 12:34:07 +0000 UTC" firstStartedPulling="2025-12-15 12:34:09.08562341 +0000 UTC m=+1010.027916440" lastFinishedPulling="2025-12-15 12:35:10.278360827 +0000 UTC m=+1071.220653867" observedRunningTime="2025-12-15 12:35:10.924046504 +0000 UTC m=+1071.866339534" watchObservedRunningTime="2025-12-15 12:35:10.926094757 +0000 UTC m=+1071.868387787" Dec 15 12:35:11 crc kubenswrapper[4719]: I1215 12:35:11.914518 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" event={"ID":"0c36396e-131f-4154-b82e-490b36d7fa63","Type":"ContainerStarted","Data":"2a16e5111088cdc499f7538aaba650325ba98be6d3b4231692d659e84cc47eb9"} Dec 15 12:35:11 crc kubenswrapper[4719]: I1215 12:35:11.939506 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jzg2t" podStartSLOduration=2.533837453 podStartE2EDuration="1m4.939489772s" podCreationTimestamp="2025-12-15 12:34:07 +0000 UTC" firstStartedPulling="2025-12-15 12:34:09.098259001 +0000 UTC m=+1010.040552021" lastFinishedPulling="2025-12-15 12:35:11.5039113 +0000 UTC m=+1072.446204340" observedRunningTime="2025-12-15 12:35:11.933823366 +0000 UTC m=+1072.876116416" watchObservedRunningTime="2025-12-15 12:35:11.939489772 +0000 UTC m=+1072.881782802" Dec 15 12:35:17 crc kubenswrapper[4719]: I1215 12:35:17.325747 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-hb66s" Dec 15 12:35:17 crc kubenswrapper[4719]: I1215 12:35:17.702929 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-l98v4" Dec 15 12:35:21 crc kubenswrapper[4719]: I1215 12:35:21.357224 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:35:21 crc kubenswrapper[4719]: I1215 12:35:21.357740 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.201579 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hgjw"] Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.203082 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.206901 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.207169 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.207371 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-tk27f" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.207514 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.222474 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hgjw"] Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.311253 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rhpjl"] Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.312623 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.323284 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.334850 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rhpjl"] Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.364810 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-config\") pod \"dnsmasq-dns-675f4bcbfc-5hgjw\" (UID: \"5a6a13b4-e84e-46db-b38d-14b4e3c827c7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.365547 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8mvb\" (UniqueName: \"kubernetes.io/projected/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-kube-api-access-h8mvb\") pod \"dnsmasq-dns-675f4bcbfc-5hgjw\" (UID: \"5a6a13b4-e84e-46db-b38d-14b4e3c827c7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.467083 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-config\") pod \"dnsmasq-dns-675f4bcbfc-5hgjw\" (UID: \"5a6a13b4-e84e-46db-b38d-14b4e3c827c7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.467156 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rhpjl\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.467188 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxz96\" (UniqueName: \"kubernetes.io/projected/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-kube-api-access-kxz96\") pod \"dnsmasq-dns-78dd6ddcc-rhpjl\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.467209 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8mvb\" (UniqueName: \"kubernetes.io/projected/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-kube-api-access-h8mvb\") pod \"dnsmasq-dns-675f4bcbfc-5hgjw\" (UID: \"5a6a13b4-e84e-46db-b38d-14b4e3c827c7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.467250 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-config\") pod \"dnsmasq-dns-78dd6ddcc-rhpjl\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.467973 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-config\") pod \"dnsmasq-dns-675f4bcbfc-5hgjw\" (UID: \"5a6a13b4-e84e-46db-b38d-14b4e3c827c7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.487164 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8mvb\" (UniqueName: \"kubernetes.io/projected/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-kube-api-access-h8mvb\") pod \"dnsmasq-dns-675f4bcbfc-5hgjw\" (UID: \"5a6a13b4-e84e-46db-b38d-14b4e3c827c7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.527896 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.568651 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-config\") pod \"dnsmasq-dns-78dd6ddcc-rhpjl\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.568992 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rhpjl\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.569110 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxz96\" (UniqueName: \"kubernetes.io/projected/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-kube-api-access-kxz96\") pod \"dnsmasq-dns-78dd6ddcc-rhpjl\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.569663 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-config\") pod \"dnsmasq-dns-78dd6ddcc-rhpjl\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.569709 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rhpjl\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.584094 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxz96\" (UniqueName: \"kubernetes.io/projected/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-kube-api-access-kxz96\") pod \"dnsmasq-dns-78dd6ddcc-rhpjl\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.631486 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:35:33 crc kubenswrapper[4719]: I1215 12:35:33.958377 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hgjw"] Dec 15 12:35:33 crc kubenswrapper[4719]: W1215 12:35:33.965796 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a6a13b4_e84e_46db_b38d_14b4e3c827c7.slice/crio-4149dc315fe8e42c0f0edc6a7cbb1afa074219209cdfac002547e690bc6e0104 WatchSource:0}: Error finding container 4149dc315fe8e42c0f0edc6a7cbb1afa074219209cdfac002547e690bc6e0104: Status 404 returned error can't find the container with id 4149dc315fe8e42c0f0edc6a7cbb1afa074219209cdfac002547e690bc6e0104 Dec 15 12:35:34 crc kubenswrapper[4719]: I1215 12:35:34.059029 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" event={"ID":"5a6a13b4-e84e-46db-b38d-14b4e3c827c7","Type":"ContainerStarted","Data":"4149dc315fe8e42c0f0edc6a7cbb1afa074219209cdfac002547e690bc6e0104"} Dec 15 12:35:34 crc kubenswrapper[4719]: I1215 12:35:34.073679 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rhpjl"] Dec 15 12:35:34 crc kubenswrapper[4719]: W1215 12:35:34.075501 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c14ae74_b1c4_494f_b0cb_ef83458cd2dc.slice/crio-ad0500275db9cfd653a01b52e14d8c3948f7c01a342510bd07a04b60217bf825 WatchSource:0}: Error finding container ad0500275db9cfd653a01b52e14d8c3948f7c01a342510bd07a04b60217bf825: Status 404 returned error can't find the container with id ad0500275db9cfd653a01b52e14d8c3948f7c01a342510bd07a04b60217bf825 Dec 15 12:35:35 crc kubenswrapper[4719]: I1215 12:35:35.068291 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" event={"ID":"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc","Type":"ContainerStarted","Data":"ad0500275db9cfd653a01b52e14d8c3948f7c01a342510bd07a04b60217bf825"} Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.638229 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hgjw"] Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.667759 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-l4qc9"] Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.668854 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.774702 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-l4qc9"] Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.859485 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpx7w\" (UniqueName: \"kubernetes.io/projected/145d5341-9595-4624-ae7b-111bc3dc0d40-kube-api-access-fpx7w\") pod \"dnsmasq-dns-666b6646f7-l4qc9\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.859542 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-config\") pod \"dnsmasq-dns-666b6646f7-l4qc9\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.859628 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-dns-svc\") pod \"dnsmasq-dns-666b6646f7-l4qc9\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.961773 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpx7w\" (UniqueName: \"kubernetes.io/projected/145d5341-9595-4624-ae7b-111bc3dc0d40-kube-api-access-fpx7w\") pod \"dnsmasq-dns-666b6646f7-l4qc9\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.961886 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-config\") pod \"dnsmasq-dns-666b6646f7-l4qc9\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.962002 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-dns-svc\") pod \"dnsmasq-dns-666b6646f7-l4qc9\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.963081 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-dns-svc\") pod \"dnsmasq-dns-666b6646f7-l4qc9\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:36 crc kubenswrapper[4719]: I1215 12:35:36.964654 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-config\") pod \"dnsmasq-dns-666b6646f7-l4qc9\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.002210 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpx7w\" (UniqueName: \"kubernetes.io/projected/145d5341-9595-4624-ae7b-111bc3dc0d40-kube-api-access-fpx7w\") pod \"dnsmasq-dns-666b6646f7-l4qc9\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.048951 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rhpjl"] Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.102368 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4mxkg"] Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.103782 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.119201 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4mxkg"] Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.266794 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd9ph\" (UniqueName: \"kubernetes.io/projected/2802b6f0-e8de-44d1-b03a-006598f261e2-kube-api-access-hd9ph\") pod \"dnsmasq-dns-57d769cc4f-4mxkg\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.267155 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-config\") pod \"dnsmasq-dns-57d769cc4f-4mxkg\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.267259 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-4mxkg\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.298900 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.369035 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-4mxkg\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.369107 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd9ph\" (UniqueName: \"kubernetes.io/projected/2802b6f0-e8de-44d1-b03a-006598f261e2-kube-api-access-hd9ph\") pod \"dnsmasq-dns-57d769cc4f-4mxkg\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.369127 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-config\") pod \"dnsmasq-dns-57d769cc4f-4mxkg\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.369849 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-config\") pod \"dnsmasq-dns-57d769cc4f-4mxkg\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.370578 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-4mxkg\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.387238 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd9ph\" (UniqueName: \"kubernetes.io/projected/2802b6f0-e8de-44d1-b03a-006598f261e2-kube-api-access-hd9ph\") pod \"dnsmasq-dns-57d769cc4f-4mxkg\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.440838 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.863254 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.864648 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.873529 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.873631 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.873811 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mwhq7" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.874106 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.874158 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.874158 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.874282 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.905614 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.981642 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.981733 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.981757 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.981777 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.981816 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-config-data\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.981844 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/62a17473-57a0-4c96-904d-a2f68bfbc92f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.981889 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxtmq\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-kube-api-access-fxtmq\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.981940 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.981970 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/62a17473-57a0-4c96-904d-a2f68bfbc92f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.981998 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:37 crc kubenswrapper[4719]: I1215 12:35:37.982021 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.029923 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-l4qc9"] Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083397 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083449 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-config-data\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083466 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/62a17473-57a0-4c96-904d-a2f68bfbc92f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083489 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxtmq\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-kube-api-access-fxtmq\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083533 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083567 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/62a17473-57a0-4c96-904d-a2f68bfbc92f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083595 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083616 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083639 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083656 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.083677 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.084563 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.085174 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.085316 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-config-data\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.085525 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.086368 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.089024 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.096690 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/62a17473-57a0-4c96-904d-a2f68bfbc92f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.097337 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.097342 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/62a17473-57a0-4c96-904d-a2f68bfbc92f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.100907 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.128959 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxtmq\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-kube-api-access-fxtmq\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.130599 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.199025 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.263720 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.266256 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.268879 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.274351 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.274536 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.274666 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.274756 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.274873 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.275584 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2df6g" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.275655 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389264 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/72f88cad-2122-46b3-9493-9f5f89890956-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389370 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389396 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkfnj\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-kube-api-access-tkfnj\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389422 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389470 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389496 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389535 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389558 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389580 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389600 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/72f88cad-2122-46b3-9493-9f5f89890956-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.389680 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.490756 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/72f88cad-2122-46b3-9493-9f5f89890956-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.491060 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.491079 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkfnj\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-kube-api-access-tkfnj\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.491099 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.491127 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.491145 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.491168 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.491186 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.491205 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.491221 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/72f88cad-2122-46b3-9493-9f5f89890956-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.491263 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.498957 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.498965 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.499140 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.499188 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.499603 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.500007 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.501067 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.504131 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/72f88cad-2122-46b3-9493-9f5f89890956-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.514610 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/72f88cad-2122-46b3-9493-9f5f89890956-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.514899 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkfnj\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-kube-api-access-tkfnj\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.565309 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.719317 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:38 crc kubenswrapper[4719]: I1215 12:35:38.896158 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.060981 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4mxkg"] Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.138119 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" event={"ID":"145d5341-9595-4624-ae7b-111bc3dc0d40","Type":"ContainerStarted","Data":"4e724f6a2157f652d4b9f1ba2afbf3b1bff0887c48eed6d89ab2bc8cade97b04"} Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.139451 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" event={"ID":"2802b6f0-e8de-44d1-b03a-006598f261e2","Type":"ContainerStarted","Data":"439c37bd9dc8b1681807b836240c0941d01eb5527d48942d7fc3710e503d3532"} Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.156924 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 15 12:35:39 crc kubenswrapper[4719]: W1215 12:35:39.163676 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62a17473_57a0_4c96_904d_a2f68bfbc92f.slice/crio-6515365ba4cd3df6a23b4ae7bdaa14e347b5356fbfa4177a18f91cc7bb4ffda9 WatchSource:0}: Error finding container 6515365ba4cd3df6a23b4ae7bdaa14e347b5356fbfa4177a18f91cc7bb4ffda9: Status 404 returned error can't find the container with id 6515365ba4cd3df6a23b4ae7bdaa14e347b5356fbfa4177a18f91cc7bb4ffda9 Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.390791 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.582616 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.600028 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.600118 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.616713 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.617245 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-r7qm7" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.617643 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.619027 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.623913 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.731573 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b614221-314c-4033-80d8-20738c1b92ae-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.731682 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3b614221-314c-4033-80d8-20738c1b92ae-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.731738 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3b614221-314c-4033-80d8-20738c1b92ae-config-data-default\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.731772 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b614221-314c-4033-80d8-20738c1b92ae-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.731803 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.731821 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b614221-314c-4033-80d8-20738c1b92ae-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.731846 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3b614221-314c-4033-80d8-20738c1b92ae-kolla-config\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.731901 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb5t5\" (UniqueName: \"kubernetes.io/projected/3b614221-314c-4033-80d8-20738c1b92ae-kube-api-access-zb5t5\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.833413 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b614221-314c-4033-80d8-20738c1b92ae-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.833472 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3b614221-314c-4033-80d8-20738c1b92ae-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.833511 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3b614221-314c-4033-80d8-20738c1b92ae-config-data-default\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.833551 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b614221-314c-4033-80d8-20738c1b92ae-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.833572 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.833588 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b614221-314c-4033-80d8-20738c1b92ae-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.833608 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3b614221-314c-4033-80d8-20738c1b92ae-kolla-config\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.833647 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb5t5\" (UniqueName: \"kubernetes.io/projected/3b614221-314c-4033-80d8-20738c1b92ae-kube-api-access-zb5t5\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.835106 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3b614221-314c-4033-80d8-20738c1b92ae-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.836474 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b614221-314c-4033-80d8-20738c1b92ae-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.836643 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.839503 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3b614221-314c-4033-80d8-20738c1b92ae-kolla-config\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.840092 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3b614221-314c-4033-80d8-20738c1b92ae-config-data-default\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.847578 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b614221-314c-4033-80d8-20738c1b92ae-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.859429 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b614221-314c-4033-80d8-20738c1b92ae-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.859984 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb5t5\" (UniqueName: \"kubernetes.io/projected/3b614221-314c-4033-80d8-20738c1b92ae-kube-api-access-zb5t5\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.866744 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3b614221-314c-4033-80d8-20738c1b92ae\") " pod="openstack/openstack-galera-0" Dec 15 12:35:39 crc kubenswrapper[4719]: I1215 12:35:39.945485 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.161883 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"62a17473-57a0-4c96-904d-a2f68bfbc92f","Type":"ContainerStarted","Data":"6515365ba4cd3df6a23b4ae7bdaa14e347b5356fbfa4177a18f91cc7bb4ffda9"} Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.163092 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"72f88cad-2122-46b3-9493-9f5f89890956","Type":"ContainerStarted","Data":"f1b3d66ec78df67cd57ae05ead32f7f4ac543fb0ead7f63c4ec0d91873a25b46"} Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.699126 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 15 12:35:40 crc kubenswrapper[4719]: W1215 12:35:40.757298 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b614221_314c_4033_80d8_20738c1b92ae.slice/crio-ddef034e6c53f18e2622f1171f7d9903f0afab417e407b231374ef00f5162e98 WatchSource:0}: Error finding container ddef034e6c53f18e2622f1171f7d9903f0afab417e407b231374ef00f5162e98: Status 404 returned error can't find the container with id ddef034e6c53f18e2622f1171f7d9903f0afab417e407b231374ef00f5162e98 Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.809402 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.810583 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.820973 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.821148 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-2zxtl" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.821336 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.821401 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.873371 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.973190 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9655f31a-6eae-461b-932a-0b1b4ca56048-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.973344 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9655f31a-6eae-461b-932a-0b1b4ca56048-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.973437 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9655f31a-6eae-461b-932a-0b1b4ca56048-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.973505 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9655f31a-6eae-461b-932a-0b1b4ca56048-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.973596 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.973646 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9655f31a-6eae-461b-932a-0b1b4ca56048-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.973694 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5wnc\" (UniqueName: \"kubernetes.io/projected/9655f31a-6eae-461b-932a-0b1b4ca56048-kube-api-access-g5wnc\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:40 crc kubenswrapper[4719]: I1215 12:35:40.973769 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655f31a-6eae-461b-932a-0b1b4ca56048-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.086003 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9655f31a-6eae-461b-932a-0b1b4ca56048-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.086080 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5wnc\" (UniqueName: \"kubernetes.io/projected/9655f31a-6eae-461b-932a-0b1b4ca56048-kube-api-access-g5wnc\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.086119 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655f31a-6eae-461b-932a-0b1b4ca56048-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.086183 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9655f31a-6eae-461b-932a-0b1b4ca56048-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.086219 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9655f31a-6eae-461b-932a-0b1b4ca56048-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.086253 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9655f31a-6eae-461b-932a-0b1b4ca56048-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.086285 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9655f31a-6eae-461b-932a-0b1b4ca56048-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.086317 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.091480 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9655f31a-6eae-461b-932a-0b1b4ca56048-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.092432 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.093023 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9655f31a-6eae-461b-932a-0b1b4ca56048-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.094980 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9655f31a-6eae-461b-932a-0b1b4ca56048-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.095844 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9655f31a-6eae-461b-932a-0b1b4ca56048-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.127912 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655f31a-6eae-461b-932a-0b1b4ca56048-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.138152 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9655f31a-6eae-461b-932a-0b1b4ca56048-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.141665 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5wnc\" (UniqueName: \"kubernetes.io/projected/9655f31a-6eae-461b-932a-0b1b4ca56048-kube-api-access-g5wnc\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.183395 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9655f31a-6eae-461b-932a-0b1b4ca56048\") " pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.211574 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3b614221-314c-4033-80d8-20738c1b92ae","Type":"ContainerStarted","Data":"ddef034e6c53f18e2622f1171f7d9903f0afab417e407b231374ef00f5162e98"} Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.286342 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.287287 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.295350 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.295536 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-7m58w" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.295680 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.321425 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.404785 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-kolla-config\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.404884 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.404908 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.404924 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-config-data\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.404961 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2294\" (UniqueName: \"kubernetes.io/projected/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-kube-api-access-v2294\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.474389 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.509430 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.509486 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.511454 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-config-data\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.511661 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-config-data\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.511793 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2294\" (UniqueName: \"kubernetes.io/projected/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-kube-api-access-v2294\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.511891 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-kolla-config\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.513392 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-kolla-config\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.521158 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.525946 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.532641 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2294\" (UniqueName: \"kubernetes.io/projected/c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1-kube-api-access-v2294\") pod \"memcached-0\" (UID: \"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1\") " pod="openstack/memcached-0" Dec 15 12:35:41 crc kubenswrapper[4719]: I1215 12:35:41.642040 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 15 12:35:42 crc kubenswrapper[4719]: I1215 12:35:42.084021 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 15 12:35:42 crc kubenswrapper[4719]: I1215 12:35:42.233632 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 15 12:35:42 crc kubenswrapper[4719]: I1215 12:35:42.271144 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9655f31a-6eae-461b-932a-0b1b4ca56048","Type":"ContainerStarted","Data":"3de8690c54e12305b81d4870fae9517b13e3a4bf782191e07142e98671f5d782"} Dec 15 12:35:43 crc kubenswrapper[4719]: I1215 12:35:43.034490 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 15 12:35:43 crc kubenswrapper[4719]: I1215 12:35:43.035365 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 15 12:35:43 crc kubenswrapper[4719]: I1215 12:35:43.041245 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f42lq\" (UniqueName: \"kubernetes.io/projected/3887591c-e8a9-43af-b2fb-f2b5cb4e1256-kube-api-access-f42lq\") pod \"kube-state-metrics-0\" (UID: \"3887591c-e8a9-43af-b2fb-f2b5cb4e1256\") " pod="openstack/kube-state-metrics-0" Dec 15 12:35:43 crc kubenswrapper[4719]: I1215 12:35:43.046377 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-95gk5" Dec 15 12:35:43 crc kubenswrapper[4719]: I1215 12:35:43.082176 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 15 12:35:43 crc kubenswrapper[4719]: I1215 12:35:43.141981 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f42lq\" (UniqueName: \"kubernetes.io/projected/3887591c-e8a9-43af-b2fb-f2b5cb4e1256-kube-api-access-f42lq\") pod \"kube-state-metrics-0\" (UID: \"3887591c-e8a9-43af-b2fb-f2b5cb4e1256\") " pod="openstack/kube-state-metrics-0" Dec 15 12:35:43 crc kubenswrapper[4719]: I1215 12:35:43.259500 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f42lq\" (UniqueName: \"kubernetes.io/projected/3887591c-e8a9-43af-b2fb-f2b5cb4e1256-kube-api-access-f42lq\") pod \"kube-state-metrics-0\" (UID: \"3887591c-e8a9-43af-b2fb-f2b5cb4e1256\") " pod="openstack/kube-state-metrics-0" Dec 15 12:35:43 crc kubenswrapper[4719]: I1215 12:35:43.328565 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1","Type":"ContainerStarted","Data":"4009b04ee4a78a933a6f6a162ad8d82028138d1d55115a7c7c761f8a2ecb3e4a"} Dec 15 12:35:43 crc kubenswrapper[4719]: I1215 12:35:43.378217 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 15 12:35:44 crc kubenswrapper[4719]: I1215 12:35:44.415398 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 15 12:35:44 crc kubenswrapper[4719]: W1215 12:35:44.478121 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3887591c_e8a9_43af_b2fb_f2b5cb4e1256.slice/crio-a4d3dc466d454757d508aeba26d9b613caca6575bc2bbf7109b587dcafed5f8c WatchSource:0}: Error finding container a4d3dc466d454757d508aeba26d9b613caca6575bc2bbf7109b587dcafed5f8c: Status 404 returned error can't find the container with id a4d3dc466d454757d508aeba26d9b613caca6575bc2bbf7109b587dcafed5f8c Dec 15 12:35:45 crc kubenswrapper[4719]: I1215 12:35:45.361372 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3887591c-e8a9-43af-b2fb-f2b5cb4e1256","Type":"ContainerStarted","Data":"a4d3dc466d454757d508aeba26d9b613caca6575bc2bbf7109b587dcafed5f8c"} Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.188208 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.191460 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.208349 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.216383 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-9lz7s" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.216975 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.217895 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.218100 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.218160 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.355506 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e0622c-a031-4927-83f3-d7d4c8dc0233-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.355546 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66e0622c-a031-4927-83f3-d7d4c8dc0233-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.355572 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/66e0622c-a031-4927-83f3-d7d4c8dc0233-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.355593 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e0622c-a031-4927-83f3-d7d4c8dc0233-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.355760 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgj4w\" (UniqueName: \"kubernetes.io/projected/66e0622c-a031-4927-83f3-d7d4c8dc0233-kube-api-access-sgj4w\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.355819 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.355889 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66e0622c-a031-4927-83f3-d7d4c8dc0233-config\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.355933 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e0622c-a031-4927-83f3-d7d4c8dc0233-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.457669 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e0622c-a031-4927-83f3-d7d4c8dc0233-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.457716 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66e0622c-a031-4927-83f3-d7d4c8dc0233-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.457743 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/66e0622c-a031-4927-83f3-d7d4c8dc0233-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.457764 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e0622c-a031-4927-83f3-d7d4c8dc0233-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.457784 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgj4w\" (UniqueName: \"kubernetes.io/projected/66e0622c-a031-4927-83f3-d7d4c8dc0233-kube-api-access-sgj4w\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.457802 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.457823 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66e0622c-a031-4927-83f3-d7d4c8dc0233-config\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.457845 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e0622c-a031-4927-83f3-d7d4c8dc0233-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.462839 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.467719 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e0622c-a031-4927-83f3-d7d4c8dc0233-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.469549 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e0622c-a031-4927-83f3-d7d4c8dc0233-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.488678 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66e0622c-a031-4927-83f3-d7d4c8dc0233-config\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.489433 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66e0622c-a031-4927-83f3-d7d4c8dc0233-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.490327 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/66e0622c-a031-4927-83f3-d7d4c8dc0233-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.497302 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e0622c-a031-4927-83f3-d7d4c8dc0233-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.499319 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgj4w\" (UniqueName: \"kubernetes.io/projected/66e0622c-a031-4927-83f3-d7d4c8dc0233-kube-api-access-sgj4w\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.528948 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"66e0622c-a031-4927-83f3-d7d4c8dc0233\") " pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:46 crc kubenswrapper[4719]: I1215 12:35:46.546289 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.584334 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4ht64"] Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.592092 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.595814 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.596139 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-mpn5r" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.597915 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ht64"] Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.600013 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.642303 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-hrrfr"] Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.656746 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.674830 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-hrrfr"] Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.711809 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7986d1-97c7-4b8c-8b2f-594fafc24991-combined-ca-bundle\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.711882 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c7986d1-97c7-4b8c-8b2f-594fafc24991-scripts\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.711933 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-etc-ovs\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.711952 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4c7986d1-97c7-4b8c-8b2f-594fafc24991-var-log-ovn\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.711978 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-var-lib\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.712020 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c7986d1-97c7-4b8c-8b2f-594fafc24991-ovn-controller-tls-certs\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.712047 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4c7986d1-97c7-4b8c-8b2f-594fafc24991-var-run\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.712091 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmdgs\" (UniqueName: \"kubernetes.io/projected/4c7986d1-97c7-4b8c-8b2f-594fafc24991-kube-api-access-xmdgs\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.712113 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-var-log\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.712136 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-var-run\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.712187 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4c7986d1-97c7-4b8c-8b2f-594fafc24991-var-run-ovn\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.712259 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcd57\" (UniqueName: \"kubernetes.io/projected/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-kube-api-access-tcd57\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.712293 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-scripts\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.813671 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7986d1-97c7-4b8c-8b2f-594fafc24991-combined-ca-bundle\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.813745 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c7986d1-97c7-4b8c-8b2f-594fafc24991-scripts\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.813778 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-etc-ovs\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.813815 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4c7986d1-97c7-4b8c-8b2f-594fafc24991-var-log-ovn\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.814174 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-var-lib\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.814212 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c7986d1-97c7-4b8c-8b2f-594fafc24991-ovn-controller-tls-certs\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.814234 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4c7986d1-97c7-4b8c-8b2f-594fafc24991-var-run\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.814256 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmdgs\" (UniqueName: \"kubernetes.io/projected/4c7986d1-97c7-4b8c-8b2f-594fafc24991-kube-api-access-xmdgs\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.814272 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-var-log\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.814315 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-var-run\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.814342 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4c7986d1-97c7-4b8c-8b2f-594fafc24991-var-run-ovn\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.814382 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcd57\" (UniqueName: \"kubernetes.io/projected/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-kube-api-access-tcd57\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.814404 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-scripts\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.817572 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4c7986d1-97c7-4b8c-8b2f-594fafc24991-var-run\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.817721 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4c7986d1-97c7-4b8c-8b2f-594fafc24991-var-run-ovn\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.817744 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-var-log\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.817774 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-var-run\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.817924 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-etc-ovs\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.817999 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4c7986d1-97c7-4b8c-8b2f-594fafc24991-var-log-ovn\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.818123 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-var-lib\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.819847 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c7986d1-97c7-4b8c-8b2f-594fafc24991-scripts\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.825584 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c7986d1-97c7-4b8c-8b2f-594fafc24991-combined-ca-bundle\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.847313 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmdgs\" (UniqueName: \"kubernetes.io/projected/4c7986d1-97c7-4b8c-8b2f-594fafc24991-kube-api-access-xmdgs\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.852693 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c7986d1-97c7-4b8c-8b2f-594fafc24991-ovn-controller-tls-certs\") pod \"ovn-controller-4ht64\" (UID: \"4c7986d1-97c7-4b8c-8b2f-594fafc24991\") " pod="openstack/ovn-controller-4ht64" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.854045 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-scripts\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.857586 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcd57\" (UniqueName: \"kubernetes.io/projected/d60bc22b-09f1-4b84-84ad-a9ee31a18a43-kube-api-access-tcd57\") pod \"ovn-controller-ovs-hrrfr\" (UID: \"d60bc22b-09f1-4b84-84ad-a9ee31a18a43\") " pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:48 crc kubenswrapper[4719]: I1215 12:35:48.921643 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ht64" Dec 15 12:35:49 crc kubenswrapper[4719]: I1215 12:35:49.005749 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.893290 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.895468 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.901328 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.901351 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.901508 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.901660 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-tjhct" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.916916 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.959544 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.959622 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.959821 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.959914 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-config\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.959956 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.959973 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.960061 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:50 crc kubenswrapper[4719]: I1215 12:35:50.960078 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6h8l\" (UniqueName: \"kubernetes.io/projected/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-kube-api-access-r6h8l\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.060840 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.061177 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.061218 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.061231 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6h8l\" (UniqueName: \"kubernetes.io/projected/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-kube-api-access-r6h8l\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.062268 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.062388 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.062433 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.062464 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-config\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.062502 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.063511 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-config\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.063973 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.064177 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.077734 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.077743 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.085623 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.087684 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6h8l\" (UniqueName: \"kubernetes.io/projected/66f93b1d-4c36-4f90-ac6c-28b302a8c34b-kube-api-access-r6h8l\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.095431 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"66f93b1d-4c36-4f90-ac6c-28b302a8c34b\") " pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.260015 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.357265 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:35:51 crc kubenswrapper[4719]: I1215 12:35:51.357333 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:36:06 crc kubenswrapper[4719]: E1215 12:36:06.583909 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 15 12:36:06 crc kubenswrapper[4719]: E1215 12:36:06.584492 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fxtmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(62a17473-57a0-4c96-904d-a2f68bfbc92f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:36:06 crc kubenswrapper[4719]: E1215 12:36:06.585810 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="62a17473-57a0-4c96-904d-a2f68bfbc92f" Dec 15 12:36:06 crc kubenswrapper[4719]: E1215 12:36:06.676835 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="62a17473-57a0-4c96-904d-a2f68bfbc92f" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.462251 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.462472 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n55ch577h556h59h589hcdh8dhcdhdh559h55dh58bh76hc4hbdh678h57ch8fh54h69h569h5bch64chfh655hc5h5d6hdch689h5f9hdfh684q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v2294,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.463768 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.473934 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.474439 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tkfnj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(72f88cad-2122-46b3-9493-9f5f89890956): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.476007 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="72f88cad-2122-46b3-9493-9f5f89890956" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.483024 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.483149 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g5wnc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(9655f31a-6eae-461b-932a-0b1b4ca56048): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.484337 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="9655f31a-6eae-461b-932a-0b1b4ca56048" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.686772 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1" Dec 15 12:36:07 crc kubenswrapper[4719]: E1215 12:36:07.690158 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="72f88cad-2122-46b3-9493-9f5f89890956" Dec 15 12:36:12 crc kubenswrapper[4719]: I1215 12:36:12.250125 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ht64"] Dec 15 12:36:12 crc kubenswrapper[4719]: I1215 12:36:12.422400 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-hrrfr"] Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.864067 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.864816 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kxz96,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-rhpjl_openstack(6c14ae74-b1c4-494f-b0cb-ef83458cd2dc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.865930 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" podUID="6c14ae74-b1c4-494f-b0cb-ef83458cd2dc" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.881204 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.881419 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hd9ph,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-4mxkg_openstack(2802b6f0-e8de-44d1-b03a-006598f261e2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.882988 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" podUID="2802b6f0-e8de-44d1-b03a-006598f261e2" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.893814 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.894300 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h8mvb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-5hgjw_openstack(5a6a13b4-e84e-46db-b38d-14b4e3c827c7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.898469 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" podUID="5a6a13b4-e84e-46db-b38d-14b4e3c827c7" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.922063 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.922498 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fpx7w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-l4qc9_openstack(145d5341-9595-4624-ae7b-111bc3dc0d40): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:36:12 crc kubenswrapper[4719]: E1215 12:36:12.923665 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" podUID="145d5341-9595-4624-ae7b-111bc3dc0d40" Dec 15 12:36:13 crc kubenswrapper[4719]: I1215 12:36:13.331200 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 15 12:36:13 crc kubenswrapper[4719]: I1215 12:36:13.443617 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 15 12:36:13 crc kubenswrapper[4719]: W1215 12:36:13.638609 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66f93b1d_4c36_4f90_ac6c_28b302a8c34b.slice/crio-728a659e9cecb4f9db5e711af004aa0b4f40a27323a06cdac1756634232fa2d4 WatchSource:0}: Error finding container 728a659e9cecb4f9db5e711af004aa0b4f40a27323a06cdac1756634232fa2d4: Status 404 returned error can't find the container with id 728a659e9cecb4f9db5e711af004aa0b4f40a27323a06cdac1756634232fa2d4 Dec 15 12:36:13 crc kubenswrapper[4719]: W1215 12:36:13.645683 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66e0622c_a031_4927_83f3_d7d4c8dc0233.slice/crio-c9c84fb34682149fa4d09f2aa901e280b8ba33a11f1da8a4bda9dc95212edcbd WatchSource:0}: Error finding container c9c84fb34682149fa4d09f2aa901e280b8ba33a11f1da8a4bda9dc95212edcbd: Status 404 returned error can't find the container with id c9c84fb34682149fa4d09f2aa901e280b8ba33a11f1da8a4bda9dc95212edcbd Dec 15 12:36:13 crc kubenswrapper[4719]: E1215 12:36:13.702119 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 15 12:36:13 crc kubenswrapper[4719]: E1215 12:36:13.702492 4719 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 15 12:36:13 crc kubenswrapper[4719]: E1215 12:36:13.704958 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f42lq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(3887591c-e8a9-43af-b2fb-f2b5cb4e1256): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 15 12:36:13 crc kubenswrapper[4719]: E1215 12:36:13.706148 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="3887591c-e8a9-43af-b2fb-f2b5cb4e1256" Dec 15 12:36:13 crc kubenswrapper[4719]: I1215 12:36:13.741812 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hrrfr" event={"ID":"d60bc22b-09f1-4b84-84ad-a9ee31a18a43","Type":"ContainerStarted","Data":"3343eb13a926ae9dcbc10cc11ef33dc8c1b9c28a88e6b3446c8646cd9b391a71"} Dec 15 12:36:13 crc kubenswrapper[4719]: I1215 12:36:13.742972 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ht64" event={"ID":"4c7986d1-97c7-4b8c-8b2f-594fafc24991","Type":"ContainerStarted","Data":"a22ac37c19e823bbff513e915819e953d05fda7f0f43b20d78db38ecc0ed2e67"} Dec 15 12:36:13 crc kubenswrapper[4719]: I1215 12:36:13.744774 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"66f93b1d-4c36-4f90-ac6c-28b302a8c34b","Type":"ContainerStarted","Data":"728a659e9cecb4f9db5e711af004aa0b4f40a27323a06cdac1756634232fa2d4"} Dec 15 12:36:13 crc kubenswrapper[4719]: I1215 12:36:13.746513 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"66e0622c-a031-4927-83f3-d7d4c8dc0233","Type":"ContainerStarted","Data":"c9c84fb34682149fa4d09f2aa901e280b8ba33a11f1da8a4bda9dc95212edcbd"} Dec 15 12:36:13 crc kubenswrapper[4719]: E1215 12:36:13.750932 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="3887591c-e8a9-43af-b2fb-f2b5cb4e1256" Dec 15 12:36:13 crc kubenswrapper[4719]: E1215 12:36:13.751095 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" podUID="2802b6f0-e8de-44d1-b03a-006598f261e2" Dec 15 12:36:13 crc kubenswrapper[4719]: E1215 12:36:13.751134 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" podUID="145d5341-9595-4624-ae7b-111bc3dc0d40" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.266398 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.271556 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.283618 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8mvb\" (UniqueName: \"kubernetes.io/projected/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-kube-api-access-h8mvb\") pod \"5a6a13b4-e84e-46db-b38d-14b4e3c827c7\" (UID: \"5a6a13b4-e84e-46db-b38d-14b4e3c827c7\") " Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.283863 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxz96\" (UniqueName: \"kubernetes.io/projected/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-kube-api-access-kxz96\") pod \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.283951 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-config\") pod \"5a6a13b4-e84e-46db-b38d-14b4e3c827c7\" (UID: \"5a6a13b4-e84e-46db-b38d-14b4e3c827c7\") " Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.283986 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-dns-svc\") pod \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.284029 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-config\") pod \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\" (UID: \"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc\") " Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.285381 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-config" (OuterVolumeSpecName: "config") pod "6c14ae74-b1c4-494f-b0cb-ef83458cd2dc" (UID: "6c14ae74-b1c4-494f-b0cb-ef83458cd2dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.286806 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-config" (OuterVolumeSpecName: "config") pod "5a6a13b4-e84e-46db-b38d-14b4e3c827c7" (UID: "5a6a13b4-e84e-46db-b38d-14b4e3c827c7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.287188 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6c14ae74-b1c4-494f-b0cb-ef83458cd2dc" (UID: "6c14ae74-b1c4-494f-b0cb-ef83458cd2dc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.294183 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-kube-api-access-h8mvb" (OuterVolumeSpecName: "kube-api-access-h8mvb") pod "5a6a13b4-e84e-46db-b38d-14b4e3c827c7" (UID: "5a6a13b4-e84e-46db-b38d-14b4e3c827c7"). InnerVolumeSpecName "kube-api-access-h8mvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.302191 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-kube-api-access-kxz96" (OuterVolumeSpecName: "kube-api-access-kxz96") pod "6c14ae74-b1c4-494f-b0cb-ef83458cd2dc" (UID: "6c14ae74-b1c4-494f-b0cb-ef83458cd2dc"). InnerVolumeSpecName "kube-api-access-kxz96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.386305 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.386360 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.386373 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.386386 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8mvb\" (UniqueName: \"kubernetes.io/projected/5a6a13b4-e84e-46db-b38d-14b4e3c827c7-kube-api-access-h8mvb\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.386539 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxz96\" (UniqueName: \"kubernetes.io/projected/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc-kube-api-access-kxz96\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.754960 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.754976 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-5hgjw" event={"ID":"5a6a13b4-e84e-46db-b38d-14b4e3c827c7","Type":"ContainerDied","Data":"4149dc315fe8e42c0f0edc6a7cbb1afa074219209cdfac002547e690bc6e0104"} Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.756780 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9655f31a-6eae-461b-932a-0b1b4ca56048","Type":"ContainerStarted","Data":"f3087778e601b099740e6df19df9439cbf24805e713e725b2f6d24726101b46f"} Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.760737 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" event={"ID":"6c14ae74-b1c4-494f-b0cb-ef83458cd2dc","Type":"ContainerDied","Data":"ad0500275db9cfd653a01b52e14d8c3948f7c01a342510bd07a04b60217bf825"} Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.760790 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rhpjl" Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.762233 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3b614221-314c-4033-80d8-20738c1b92ae","Type":"ContainerStarted","Data":"02406d76c9c12c0f4c4a065dc5925d8075ebd8146688995c978a759a575d1d6e"} Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.874374 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hgjw"] Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.880614 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-5hgjw"] Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.890776 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rhpjl"] Dec 15 12:36:14 crc kubenswrapper[4719]: I1215 12:36:14.919577 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rhpjl"] Dec 15 12:36:15 crc kubenswrapper[4719]: I1215 12:36:15.484635 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a6a13b4-e84e-46db-b38d-14b4e3c827c7" path="/var/lib/kubelet/pods/5a6a13b4-e84e-46db-b38d-14b4e3c827c7/volumes" Dec 15 12:36:15 crc kubenswrapper[4719]: I1215 12:36:15.485046 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c14ae74-b1c4-494f-b0cb-ef83458cd2dc" path="/var/lib/kubelet/pods/6c14ae74-b1c4-494f-b0cb-ef83458cd2dc/volumes" Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.808424 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"66f93b1d-4c36-4f90-ac6c-28b302a8c34b","Type":"ContainerStarted","Data":"daf71bc1a68c19da484abaeda44c3f513532bb5a5d4c321f7047efcbb9388192"} Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.811165 4719 generic.go:334] "Generic (PLEG): container finished" podID="9655f31a-6eae-461b-932a-0b1b4ca56048" containerID="f3087778e601b099740e6df19df9439cbf24805e713e725b2f6d24726101b46f" exitCode=0 Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.811260 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9655f31a-6eae-461b-932a-0b1b4ca56048","Type":"ContainerDied","Data":"f3087778e601b099740e6df19df9439cbf24805e713e725b2f6d24726101b46f"} Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.815077 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"66e0622c-a031-4927-83f3-d7d4c8dc0233","Type":"ContainerStarted","Data":"0304343b094de38add91485a34d9aa994f4073dde08c2d1ea3898c8b51e6d89a"} Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.821144 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hrrfr" event={"ID":"d60bc22b-09f1-4b84-84ad-a9ee31a18a43","Type":"ContainerStarted","Data":"e6ea3a7f4425e8e530b6d5584790e9bc16386bd8d2003557e04670130d9eb69c"} Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.823208 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ht64" event={"ID":"4c7986d1-97c7-4b8c-8b2f-594fafc24991","Type":"ContainerStarted","Data":"a43ef6eb845022fff01b842421bcedb6094daf897eb8a9a5282a621220ebecac"} Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.823401 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-4ht64" Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.824787 4719 generic.go:334] "Generic (PLEG): container finished" podID="3b614221-314c-4033-80d8-20738c1b92ae" containerID="02406d76c9c12c0f4c4a065dc5925d8075ebd8146688995c978a759a575d1d6e" exitCode=0 Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.824820 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3b614221-314c-4033-80d8-20738c1b92ae","Type":"ContainerDied","Data":"02406d76c9c12c0f4c4a065dc5925d8075ebd8146688995c978a759a575d1d6e"} Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.912525 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-84b56"] Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.913662 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.919038 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.921635 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4ht64" podStartSLOduration=25.490764317 podStartE2EDuration="32.921621579s" podCreationTimestamp="2025-12-15 12:35:48 +0000 UTC" firstStartedPulling="2025-12-15 12:36:12.838038726 +0000 UTC m=+1133.780331796" lastFinishedPulling="2025-12-15 12:36:20.268895998 +0000 UTC m=+1141.211189058" observedRunningTime="2025-12-15 12:36:20.916910871 +0000 UTC m=+1141.859203921" watchObservedRunningTime="2025-12-15 12:36:20.921621579 +0000 UTC m=+1141.863914609" Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.928353 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-84b56"] Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.990152 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dc6d84f-343b-456a-bad8-3171dccc2b9d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.990221 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dc6d84f-343b-456a-bad8-3171dccc2b9d-combined-ca-bundle\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.990247 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8dc6d84f-343b-456a-bad8-3171dccc2b9d-ovs-rundir\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.990293 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dc6d84f-343b-456a-bad8-3171dccc2b9d-config\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.990446 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcfsj\" (UniqueName: \"kubernetes.io/projected/8dc6d84f-343b-456a-bad8-3171dccc2b9d-kube-api-access-kcfsj\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:20 crc kubenswrapper[4719]: I1215 12:36:20.990513 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8dc6d84f-343b-456a-bad8-3171dccc2b9d-ovn-rundir\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.089800 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4mxkg"] Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.092164 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dc6d84f-343b-456a-bad8-3171dccc2b9d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.095233 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dc6d84f-343b-456a-bad8-3171dccc2b9d-combined-ca-bundle\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.095317 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8dc6d84f-343b-456a-bad8-3171dccc2b9d-ovs-rundir\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.095370 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dc6d84f-343b-456a-bad8-3171dccc2b9d-config\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.095550 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcfsj\" (UniqueName: \"kubernetes.io/projected/8dc6d84f-343b-456a-bad8-3171dccc2b9d-kube-api-access-kcfsj\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.095626 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8dc6d84f-343b-456a-bad8-3171dccc2b9d-ovn-rundir\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.095975 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8dc6d84f-343b-456a-bad8-3171dccc2b9d-ovn-rundir\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.096025 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8dc6d84f-343b-456a-bad8-3171dccc2b9d-ovs-rundir\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.096653 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dc6d84f-343b-456a-bad8-3171dccc2b9d-config\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.147811 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dc6d84f-343b-456a-bad8-3171dccc2b9d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.148177 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dc6d84f-343b-456a-bad8-3171dccc2b9d-combined-ca-bundle\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.165210 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcfsj\" (UniqueName: \"kubernetes.io/projected/8dc6d84f-343b-456a-bad8-3171dccc2b9d-kube-api-access-kcfsj\") pod \"ovn-controller-metrics-84b56\" (UID: \"8dc6d84f-343b-456a-bad8-3171dccc2b9d\") " pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.173035 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f6tsl"] Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.174267 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f6tsl"] Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.174350 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.182543 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.260044 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-84b56" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.287849 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-l4qc9"] Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.305497 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.305579 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-config\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.305628 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcmms\" (UniqueName: \"kubernetes.io/projected/15c6924c-50fe-418b-bbde-f890493f82ed-kube-api-access-gcmms\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.305652 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.322334 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-lnm79"] Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.338562 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.342742 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.353531 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-lnm79"] Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.357006 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.357057 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.357104 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.357773 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6512c03468894d57c0ec7081d538364858ba849e9d66b22fa695a25587406622"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.357831 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://6512c03468894d57c0ec7081d538364858ba849e9d66b22fa695a25587406622" gracePeriod=600 Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.409994 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.410044 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc6mf\" (UniqueName: \"kubernetes.io/projected/aa6437c9-add6-4169-b032-406f5ddfd4e6-kube-api-access-fc6mf\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.410101 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-config\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.410119 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-config\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.410169 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.410186 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.410202 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcmms\" (UniqueName: \"kubernetes.io/projected/15c6924c-50fe-418b-bbde-f890493f82ed-kube-api-access-gcmms\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.410223 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.410247 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.411332 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-config\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.415551 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.416261 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.437176 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcmms\" (UniqueName: \"kubernetes.io/projected/15c6924c-50fe-418b-bbde-f890493f82ed-kube-api-access-gcmms\") pod \"dnsmasq-dns-7fd796d7df-f6tsl\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.525331 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc6mf\" (UniqueName: \"kubernetes.io/projected/aa6437c9-add6-4169-b032-406f5ddfd4e6-kube-api-access-fc6mf\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.525436 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-config\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.525498 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.525520 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.525560 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.526105 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.529731 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.530107 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-config\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.530554 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.530578 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.581651 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc6mf\" (UniqueName: \"kubernetes.io/projected/aa6437c9-add6-4169-b032-406f5ddfd4e6-kube-api-access-fc6mf\") pod \"dnsmasq-dns-86db49b7ff-lnm79\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.829278 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.832440 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.835157 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3b614221-314c-4033-80d8-20738c1b92ae","Type":"ContainerStarted","Data":"371d773e038517284529c7755d7cf221d13c0f4ae9518bafb7451ca81dd0576c"} Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.838668 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="6512c03468894d57c0ec7081d538364858ba849e9d66b22fa695a25587406622" exitCode=0 Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.838722 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"6512c03468894d57c0ec7081d538364858ba849e9d66b22fa695a25587406622"} Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.838754 4719 scope.go:117] "RemoveContainer" containerID="2d86014a7f0b55979e42be548b06831288da0c2f9ca42323013a65f25c4ad994" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.847342 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" event={"ID":"2802b6f0-e8de-44d1-b03a-006598f261e2","Type":"ContainerDied","Data":"439c37bd9dc8b1681807b836240c0941d01eb5527d48942d7fc3710e503d3532"} Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.847459 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4mxkg" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.859941 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" event={"ID":"145d5341-9595-4624-ae7b-111bc3dc0d40","Type":"ContainerDied","Data":"4e724f6a2157f652d4b9f1ba2afbf3b1bff0887c48eed6d89ab2bc8cade97b04"} Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.859975 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e724f6a2157f652d4b9f1ba2afbf3b1bff0887c48eed6d89ab2bc8cade97b04" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.889825 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f6tsl"] Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.896668 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=12.898901311 podStartE2EDuration="43.896652656s" podCreationTimestamp="2025-12-15 12:35:38 +0000 UTC" firstStartedPulling="2025-12-15 12:35:40.812102151 +0000 UTC m=+1101.754395191" lastFinishedPulling="2025-12-15 12:36:11.809853496 +0000 UTC m=+1132.752146536" observedRunningTime="2025-12-15 12:36:21.874620377 +0000 UTC m=+1142.816913407" watchObservedRunningTime="2025-12-15 12:36:21.896652656 +0000 UTC m=+1142.838945686" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.934601 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd9ph\" (UniqueName: \"kubernetes.io/projected/2802b6f0-e8de-44d1-b03a-006598f261e2-kube-api-access-hd9ph\") pod \"2802b6f0-e8de-44d1-b03a-006598f261e2\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.934726 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-config\") pod \"2802b6f0-e8de-44d1-b03a-006598f261e2\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.934786 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-dns-svc\") pod \"2802b6f0-e8de-44d1-b03a-006598f261e2\" (UID: \"2802b6f0-e8de-44d1-b03a-006598f261e2\") " Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.936138 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-config" (OuterVolumeSpecName: "config") pod "2802b6f0-e8de-44d1-b03a-006598f261e2" (UID: "2802b6f0-e8de-44d1-b03a-006598f261e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.936524 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2802b6f0-e8de-44d1-b03a-006598f261e2" (UID: "2802b6f0-e8de-44d1-b03a-006598f261e2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.940579 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.949058 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2802b6f0-e8de-44d1-b03a-006598f261e2-kube-api-access-hd9ph" (OuterVolumeSpecName: "kube-api-access-hd9ph") pod "2802b6f0-e8de-44d1-b03a-006598f261e2" (UID: "2802b6f0-e8de-44d1-b03a-006598f261e2"). InnerVolumeSpecName "kube-api-access-hd9ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:21 crc kubenswrapper[4719]: I1215 12:36:21.979270 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-84b56"] Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.037749 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-config\") pod \"145d5341-9595-4624-ae7b-111bc3dc0d40\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.038143 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpx7w\" (UniqueName: \"kubernetes.io/projected/145d5341-9595-4624-ae7b-111bc3dc0d40-kube-api-access-fpx7w\") pod \"145d5341-9595-4624-ae7b-111bc3dc0d40\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.038179 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-dns-svc\") pod \"145d5341-9595-4624-ae7b-111bc3dc0d40\" (UID: \"145d5341-9595-4624-ae7b-111bc3dc0d40\") " Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.038645 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.038661 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd9ph\" (UniqueName: \"kubernetes.io/projected/2802b6f0-e8de-44d1-b03a-006598f261e2-kube-api-access-hd9ph\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.038676 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2802b6f0-e8de-44d1-b03a-006598f261e2-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.039339 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-config" (OuterVolumeSpecName: "config") pod "145d5341-9595-4624-ae7b-111bc3dc0d40" (UID: "145d5341-9595-4624-ae7b-111bc3dc0d40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.039527 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "145d5341-9595-4624-ae7b-111bc3dc0d40" (UID: "145d5341-9595-4624-ae7b-111bc3dc0d40"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.047988 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/145d5341-9595-4624-ae7b-111bc3dc0d40-kube-api-access-fpx7w" (OuterVolumeSpecName: "kube-api-access-fpx7w") pod "145d5341-9595-4624-ae7b-111bc3dc0d40" (UID: "145d5341-9595-4624-ae7b-111bc3dc0d40"). InnerVolumeSpecName "kube-api-access-fpx7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.140170 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.140203 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpx7w\" (UniqueName: \"kubernetes.io/projected/145d5341-9595-4624-ae7b-111bc3dc0d40-kube-api-access-fpx7w\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.140214 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/145d5341-9595-4624-ae7b-111bc3dc0d40-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.218574 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4mxkg"] Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.228419 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4mxkg"] Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.313696 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-lnm79"] Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.870894 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"62a17473-57a0-4c96-904d-a2f68bfbc92f","Type":"ContainerStarted","Data":"0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7"} Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.873617 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" event={"ID":"15c6924c-50fe-418b-bbde-f890493f82ed","Type":"ContainerStarted","Data":"22c66f14ccca7c4873e9a034cb36a314adcb8f7385df742fc58548892dc2fc31"} Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.876000 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9655f31a-6eae-461b-932a-0b1b4ca56048","Type":"ContainerStarted","Data":"5dc88bc619932a326838ee140613c8ae8d114c362288d1ea9a1cbe9633aadc15"} Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.879597 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"72f88cad-2122-46b3-9493-9f5f89890956","Type":"ContainerStarted","Data":"6e34719f6e20b66097cb87d45428b9ef0d5203fb97ac81e566f14c57bc6c4811"} Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.885438 4719 generic.go:334] "Generic (PLEG): container finished" podID="d60bc22b-09f1-4b84-84ad-a9ee31a18a43" containerID="e6ea3a7f4425e8e530b6d5584790e9bc16386bd8d2003557e04670130d9eb69c" exitCode=0 Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.885583 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hrrfr" event={"ID":"d60bc22b-09f1-4b84-84ad-a9ee31a18a43","Type":"ContainerDied","Data":"e6ea3a7f4425e8e530b6d5584790e9bc16386bd8d2003557e04670130d9eb69c"} Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.886951 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-84b56" event={"ID":"8dc6d84f-343b-456a-bad8-3171dccc2b9d","Type":"ContainerStarted","Data":"4fbec03146fb0a63abc6696da0456afe3592bf7252da77cb4e727f54bbd04e87"} Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.888764 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" event={"ID":"aa6437c9-add6-4169-b032-406f5ddfd4e6","Type":"ContainerStarted","Data":"ac465cc236861554b8752c28506eff32f53e45c46bf932566578f422424e48a1"} Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.892803 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"e82d238054920207f8e0cb0360298c4a0049bf10cc2919ab754c599ae7031178"} Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.893009 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-l4qc9" Dec 15 12:36:22 crc kubenswrapper[4719]: I1215 12:36:22.976137 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371992.878674 podStartE2EDuration="43.976101419s" podCreationTimestamp="2025-12-15 12:35:39 +0000 UTC" firstStartedPulling="2025-12-15 12:35:42.12187833 +0000 UTC m=+1103.064171360" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:36:22.947333389 +0000 UTC m=+1143.889626439" watchObservedRunningTime="2025-12-15 12:36:22.976101419 +0000 UTC m=+1143.918394449" Dec 15 12:36:23 crc kubenswrapper[4719]: I1215 12:36:23.034976 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-l4qc9"] Dec 15 12:36:23 crc kubenswrapper[4719]: I1215 12:36:23.047569 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-l4qc9"] Dec 15 12:36:23 crc kubenswrapper[4719]: I1215 12:36:23.490002 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="145d5341-9595-4624-ae7b-111bc3dc0d40" path="/var/lib/kubelet/pods/145d5341-9595-4624-ae7b-111bc3dc0d40/volumes" Dec 15 12:36:23 crc kubenswrapper[4719]: I1215 12:36:23.490538 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2802b6f0-e8de-44d1-b03a-006598f261e2" path="/var/lib/kubelet/pods/2802b6f0-e8de-44d1-b03a-006598f261e2/volumes" Dec 15 12:36:26 crc kubenswrapper[4719]: I1215 12:36:26.920694 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hrrfr" event={"ID":"d60bc22b-09f1-4b84-84ad-a9ee31a18a43","Type":"ContainerStarted","Data":"e0618764da86633cbdbf405904a18b11b8cbb9b1305137af377b4066a3e0b269"} Dec 15 12:36:27 crc kubenswrapper[4719]: E1215 12:36:27.106333 4719 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.217:58208->38.102.83.217:45485: write tcp 38.102.83.217:58208->38.102.83.217:45485: write: broken pipe Dec 15 12:36:27 crc kubenswrapper[4719]: E1215 12:36:27.258462 4719 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.217:58222->38.102.83.217:45485: write tcp 38.102.83.217:58222->38.102.83.217:45485: write: connection reset by peer Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.935206 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"66f93b1d-4c36-4f90-ac6c-28b302a8c34b","Type":"ContainerStarted","Data":"69bc86644d5a4496a7466c10b02adb4403e0bebfc1978dfdb094704fec21dd95"} Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.937277 4719 generic.go:334] "Generic (PLEG): container finished" podID="15c6924c-50fe-418b-bbde-f890493f82ed" containerID="9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e" exitCode=0 Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.937333 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" event={"ID":"15c6924c-50fe-418b-bbde-f890493f82ed","Type":"ContainerDied","Data":"9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e"} Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.940505 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1","Type":"ContainerStarted","Data":"b33264dfc300826985bbd2ac2d93b008a7bdda6c407696aa083ff9bcf689d226"} Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.941561 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.946970 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"66e0622c-a031-4927-83f3-d7d4c8dc0233","Type":"ContainerStarted","Data":"dc89a4c80360f4954199bf547e97f9d125c278343fdc6ffe1457122966d21bf7"} Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.949354 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-hrrfr" event={"ID":"d60bc22b-09f1-4b84-84ad-a9ee31a18a43","Type":"ContainerStarted","Data":"3e09b287cf95559957584a5e49102d20354e0eb86a10ed34d5e04e0f24ec57ec"} Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.949925 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.949951 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.950968 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3887591c-e8a9-43af-b2fb-f2b5cb4e1256","Type":"ContainerStarted","Data":"9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b"} Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.951565 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.952599 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-84b56" event={"ID":"8dc6d84f-343b-456a-bad8-3171dccc2b9d","Type":"ContainerStarted","Data":"5abc6877ecf7024d6d5c45e68cec1fc0e970c48d20021e5638dab989d9259363"} Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.954165 4719 generic.go:334] "Generic (PLEG): container finished" podID="aa6437c9-add6-4169-b032-406f5ddfd4e6" containerID="d4ab65f7f3ef39846709e553b53cbbbaf4e4a5de4fa8444b49c4af6eae027c3d" exitCode=0 Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.954195 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" event={"ID":"aa6437c9-add6-4169-b032-406f5ddfd4e6","Type":"ContainerDied","Data":"d4ab65f7f3ef39846709e553b53cbbbaf4e4a5de4fa8444b49c4af6eae027c3d"} Dec 15 12:36:28 crc kubenswrapper[4719]: I1215 12:36:28.987987 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=25.219693869 podStartE2EDuration="39.987962072s" podCreationTimestamp="2025-12-15 12:35:49 +0000 UTC" firstStartedPulling="2025-12-15 12:36:13.641452798 +0000 UTC m=+1134.583745828" lastFinishedPulling="2025-12-15 12:36:28.409721001 +0000 UTC m=+1149.352014031" observedRunningTime="2025-12-15 12:36:28.965134519 +0000 UTC m=+1149.907427549" watchObservedRunningTime="2025-12-15 12:36:28.987962072 +0000 UTC m=+1149.930255122" Dec 15 12:36:29 crc kubenswrapper[4719]: I1215 12:36:29.068602 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.215223315 podStartE2EDuration="46.068580353s" podCreationTimestamp="2025-12-15 12:35:43 +0000 UTC" firstStartedPulling="2025-12-15 12:35:44.485967522 +0000 UTC m=+1105.428260552" lastFinishedPulling="2025-12-15 12:36:28.33932455 +0000 UTC m=+1149.281617590" observedRunningTime="2025-12-15 12:36:28.997139489 +0000 UTC m=+1149.939432519" watchObservedRunningTime="2025-12-15 12:36:29.068580353 +0000 UTC m=+1150.010873383" Dec 15 12:36:29 crc kubenswrapper[4719]: I1215 12:36:29.070783 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=29.371871468 podStartE2EDuration="44.070772522s" podCreationTimestamp="2025-12-15 12:35:45 +0000 UTC" firstStartedPulling="2025-12-15 12:36:13.665848561 +0000 UTC m=+1134.608141591" lastFinishedPulling="2025-12-15 12:36:28.364749625 +0000 UTC m=+1149.307042645" observedRunningTime="2025-12-15 12:36:29.034007392 +0000 UTC m=+1149.976300422" watchObservedRunningTime="2025-12-15 12:36:29.070772522 +0000 UTC m=+1150.013065562" Dec 15 12:36:29 crc kubenswrapper[4719]: I1215 12:36:29.138478 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=4.041924607 podStartE2EDuration="48.138458398s" podCreationTimestamp="2025-12-15 12:35:41 +0000 UTC" firstStartedPulling="2025-12-15 12:35:42.271152878 +0000 UTC m=+1103.213445918" lastFinishedPulling="2025-12-15 12:36:26.367686679 +0000 UTC m=+1147.309979709" observedRunningTime="2025-12-15 12:36:29.08253185 +0000 UTC m=+1150.024824890" watchObservedRunningTime="2025-12-15 12:36:29.138458398 +0000 UTC m=+1150.080751428" Dec 15 12:36:29 crc kubenswrapper[4719]: I1215 12:36:29.151628 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-hrrfr" podStartSLOduration=33.724397132 podStartE2EDuration="41.151585109s" podCreationTimestamp="2025-12-15 12:35:48 +0000 UTC" firstStartedPulling="2025-12-15 12:36:12.841768013 +0000 UTC m=+1133.784061043" lastFinishedPulling="2025-12-15 12:36:20.268956 +0000 UTC m=+1141.211249020" observedRunningTime="2025-12-15 12:36:29.114255681 +0000 UTC m=+1150.056548711" watchObservedRunningTime="2025-12-15 12:36:29.151585109 +0000 UTC m=+1150.093878149" Dec 15 12:36:29 crc kubenswrapper[4719]: I1215 12:36:29.253482 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-84b56" podStartSLOduration=2.858200703 podStartE2EDuration="9.253457154s" podCreationTimestamp="2025-12-15 12:36:20 +0000 UTC" firstStartedPulling="2025-12-15 12:36:21.995193137 +0000 UTC m=+1142.937486167" lastFinishedPulling="2025-12-15 12:36:28.390449588 +0000 UTC m=+1149.332742618" observedRunningTime="2025-12-15 12:36:29.172890244 +0000 UTC m=+1150.115183274" watchObservedRunningTime="2025-12-15 12:36:29.253457154 +0000 UTC m=+1150.195750184" Dec 15 12:36:29 crc kubenswrapper[4719]: I1215 12:36:29.946374 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 15 12:36:29 crc kubenswrapper[4719]: I1215 12:36:29.947772 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 15 12:36:30 crc kubenswrapper[4719]: I1215 12:36:30.193050 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 15 12:36:30 crc kubenswrapper[4719]: I1215 12:36:30.261895 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 15 12:36:30 crc kubenswrapper[4719]: I1215 12:36:30.262676 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 15 12:36:30 crc kubenswrapper[4719]: I1215 12:36:30.319494 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 15 12:36:30 crc kubenswrapper[4719]: I1215 12:36:30.967267 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" event={"ID":"aa6437c9-add6-4169-b032-406f5ddfd4e6","Type":"ContainerStarted","Data":"1ab918d28fd7377927cac0863fae8fee302336c33169df244f368a1bd099f2fe"} Dec 15 12:36:30 crc kubenswrapper[4719]: I1215 12:36:30.967416 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:30 crc kubenswrapper[4719]: I1215 12:36:30.968918 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" event={"ID":"15c6924c-50fe-418b-bbde-f890493f82ed","Type":"ContainerStarted","Data":"d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030"} Dec 15 12:36:30 crc kubenswrapper[4719]: I1215 12:36:30.969470 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 15 12:36:30 crc kubenswrapper[4719]: I1215 12:36:30.993638 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" podStartSLOduration=4.088139071 podStartE2EDuration="9.993603477s" podCreationTimestamp="2025-12-15 12:36:21 +0000 UTC" firstStartedPulling="2025-12-15 12:36:22.327977583 +0000 UTC m=+1143.270270613" lastFinishedPulling="2025-12-15 12:36:28.233441989 +0000 UTC m=+1149.175735019" observedRunningTime="2025-12-15 12:36:30.991159 +0000 UTC m=+1151.933452030" watchObservedRunningTime="2025-12-15 12:36:30.993603477 +0000 UTC m=+1151.935896557" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.009493 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.021338 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" podStartSLOduration=3.56143299 podStartE2EDuration="10.021317153s" podCreationTimestamp="2025-12-15 12:36:21 +0000 UTC" firstStartedPulling="2025-12-15 12:36:21.930566576 +0000 UTC m=+1142.872859606" lastFinishedPulling="2025-12-15 12:36:28.390450739 +0000 UTC m=+1149.332743769" observedRunningTime="2025-12-15 12:36:31.019981652 +0000 UTC m=+1151.962274692" watchObservedRunningTime="2025-12-15 12:36:31.021317153 +0000 UTC m=+1151.963610203" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.252030 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-9mxg6"] Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.254571 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9mxg6" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.266124 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2498-account-create-update-bkr7v"] Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.267443 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2498-account-create-update-bkr7v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.272027 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2498-account-create-update-bkr7v"] Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.276946 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.278537 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9mxg6"] Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.355113 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8ghm\" (UniqueName: \"kubernetes.io/projected/392caf51-84a6-43d0-9bae-edbf23c1ef0a-kube-api-access-t8ghm\") pod \"keystone-2498-account-create-update-bkr7v\" (UID: \"392caf51-84a6-43d0-9bae-edbf23c1ef0a\") " pod="openstack/keystone-2498-account-create-update-bkr7v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.355200 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/392caf51-84a6-43d0-9bae-edbf23c1ef0a-operator-scripts\") pod \"keystone-2498-account-create-update-bkr7v\" (UID: \"392caf51-84a6-43d0-9bae-edbf23c1ef0a\") " pod="openstack/keystone-2498-account-create-update-bkr7v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.355232 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpkcl\" (UniqueName: \"kubernetes.io/projected/709db400-dea1-4f84-a1bd-e3c6bdc44644-kube-api-access-qpkcl\") pod \"keystone-db-create-9mxg6\" (UID: \"709db400-dea1-4f84-a1bd-e3c6bdc44644\") " pod="openstack/keystone-db-create-9mxg6" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.355412 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/709db400-dea1-4f84-a1bd-e3c6bdc44644-operator-scripts\") pod \"keystone-db-create-9mxg6\" (UID: \"709db400-dea1-4f84-a1bd-e3c6bdc44644\") " pod="openstack/keystone-db-create-9mxg6" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.457263 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/709db400-dea1-4f84-a1bd-e3c6bdc44644-operator-scripts\") pod \"keystone-db-create-9mxg6\" (UID: \"709db400-dea1-4f84-a1bd-e3c6bdc44644\") " pod="openstack/keystone-db-create-9mxg6" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.457382 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8ghm\" (UniqueName: \"kubernetes.io/projected/392caf51-84a6-43d0-9bae-edbf23c1ef0a-kube-api-access-t8ghm\") pod \"keystone-2498-account-create-update-bkr7v\" (UID: \"392caf51-84a6-43d0-9bae-edbf23c1ef0a\") " pod="openstack/keystone-2498-account-create-update-bkr7v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.457485 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/392caf51-84a6-43d0-9bae-edbf23c1ef0a-operator-scripts\") pod \"keystone-2498-account-create-update-bkr7v\" (UID: \"392caf51-84a6-43d0-9bae-edbf23c1ef0a\") " pod="openstack/keystone-2498-account-create-update-bkr7v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.457519 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpkcl\" (UniqueName: \"kubernetes.io/projected/709db400-dea1-4f84-a1bd-e3c6bdc44644-kube-api-access-qpkcl\") pod \"keystone-db-create-9mxg6\" (UID: \"709db400-dea1-4f84-a1bd-e3c6bdc44644\") " pod="openstack/keystone-db-create-9mxg6" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.458919 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/709db400-dea1-4f84-a1bd-e3c6bdc44644-operator-scripts\") pod \"keystone-db-create-9mxg6\" (UID: \"709db400-dea1-4f84-a1bd-e3c6bdc44644\") " pod="openstack/keystone-db-create-9mxg6" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.459609 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/392caf51-84a6-43d0-9bae-edbf23c1ef0a-operator-scripts\") pod \"keystone-2498-account-create-update-bkr7v\" (UID: \"392caf51-84a6-43d0-9bae-edbf23c1ef0a\") " pod="openstack/keystone-2498-account-create-update-bkr7v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.475047 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.475190 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.484822 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8ghm\" (UniqueName: \"kubernetes.io/projected/392caf51-84a6-43d0-9bae-edbf23c1ef0a-kube-api-access-t8ghm\") pod \"keystone-2498-account-create-update-bkr7v\" (UID: \"392caf51-84a6-43d0-9bae-edbf23c1ef0a\") " pod="openstack/keystone-2498-account-create-update-bkr7v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.487732 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpkcl\" (UniqueName: \"kubernetes.io/projected/709db400-dea1-4f84-a1bd-e3c6bdc44644-kube-api-access-qpkcl\") pod \"keystone-db-create-9mxg6\" (UID: \"709db400-dea1-4f84-a1bd-e3c6bdc44644\") " pod="openstack/keystone-db-create-9mxg6" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.508401 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-5jgq2"] Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.510637 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5jgq2" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.523336 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5jgq2"] Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.527001 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.547138 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.547191 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.558824 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5121ea0-ad37-471e-9eca-d88c97375e25-operator-scripts\") pod \"placement-db-create-5jgq2\" (UID: \"e5121ea0-ad37-471e-9eca-d88c97375e25\") " pod="openstack/placement-db-create-5jgq2" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.558931 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrzwd\" (UniqueName: \"kubernetes.io/projected/e5121ea0-ad37-471e-9eca-d88c97375e25-kube-api-access-mrzwd\") pod \"placement-db-create-5jgq2\" (UID: \"e5121ea0-ad37-471e-9eca-d88c97375e25\") " pod="openstack/placement-db-create-5jgq2" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.581871 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9mxg6" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.586819 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.590874 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2498-account-create-update-bkr7v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.594749 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.647029 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-00d1-account-create-update-tpf9v"] Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.648124 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-00d1-account-create-update-tpf9v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.655110 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.662274 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrzwd\" (UniqueName: \"kubernetes.io/projected/e5121ea0-ad37-471e-9eca-d88c97375e25-kube-api-access-mrzwd\") pod \"placement-db-create-5jgq2\" (UID: \"e5121ea0-ad37-471e-9eca-d88c97375e25\") " pod="openstack/placement-db-create-5jgq2" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.662445 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5121ea0-ad37-471e-9eca-d88c97375e25-operator-scripts\") pod \"placement-db-create-5jgq2\" (UID: \"e5121ea0-ad37-471e-9eca-d88c97375e25\") " pod="openstack/placement-db-create-5jgq2" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.663287 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5121ea0-ad37-471e-9eca-d88c97375e25-operator-scripts\") pod \"placement-db-create-5jgq2\" (UID: \"e5121ea0-ad37-471e-9eca-d88c97375e25\") " pod="openstack/placement-db-create-5jgq2" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.678998 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-00d1-account-create-update-tpf9v"] Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.685022 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrzwd\" (UniqueName: \"kubernetes.io/projected/e5121ea0-ad37-471e-9eca-d88c97375e25-kube-api-access-mrzwd\") pod \"placement-db-create-5jgq2\" (UID: \"e5121ea0-ad37-471e-9eca-d88c97375e25\") " pod="openstack/placement-db-create-5jgq2" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.767680 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxr5h\" (UniqueName: \"kubernetes.io/projected/50261891-12d4-4ec9-a0c0-e3b1d07859b6-kube-api-access-sxr5h\") pod \"placement-00d1-account-create-update-tpf9v\" (UID: \"50261891-12d4-4ec9-a0c0-e3b1d07859b6\") " pod="openstack/placement-00d1-account-create-update-tpf9v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.769958 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50261891-12d4-4ec9-a0c0-e3b1d07859b6-operator-scripts\") pod \"placement-00d1-account-create-update-tpf9v\" (UID: \"50261891-12d4-4ec9-a0c0-e3b1d07859b6\") " pod="openstack/placement-00d1-account-create-update-tpf9v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.869723 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5jgq2" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.871673 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50261891-12d4-4ec9-a0c0-e3b1d07859b6-operator-scripts\") pod \"placement-00d1-account-create-update-tpf9v\" (UID: \"50261891-12d4-4ec9-a0c0-e3b1d07859b6\") " pod="openstack/placement-00d1-account-create-update-tpf9v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.871763 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxr5h\" (UniqueName: \"kubernetes.io/projected/50261891-12d4-4ec9-a0c0-e3b1d07859b6-kube-api-access-sxr5h\") pod \"placement-00d1-account-create-update-tpf9v\" (UID: \"50261891-12d4-4ec9-a0c0-e3b1d07859b6\") " pod="openstack/placement-00d1-account-create-update-tpf9v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.872359 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50261891-12d4-4ec9-a0c0-e3b1d07859b6-operator-scripts\") pod \"placement-00d1-account-create-update-tpf9v\" (UID: \"50261891-12d4-4ec9-a0c0-e3b1d07859b6\") " pod="openstack/placement-00d1-account-create-update-tpf9v" Dec 15 12:36:31 crc kubenswrapper[4719]: I1215 12:36:31.887728 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxr5h\" (UniqueName: \"kubernetes.io/projected/50261891-12d4-4ec9-a0c0-e3b1d07859b6-kube-api-access-sxr5h\") pod \"placement-00d1-account-create-update-tpf9v\" (UID: \"50261891-12d4-4ec9-a0c0-e3b1d07859b6\") " pod="openstack/placement-00d1-account-create-update-tpf9v" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.036479 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.043481 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-00d1-account-create-update-tpf9v" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.091793 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.127182 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9mxg6"] Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.178875 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2498-account-create-update-bkr7v"] Dec 15 12:36:32 crc kubenswrapper[4719]: W1215 12:36:32.285065 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod392caf51_84a6_43d0_9bae_edbf23c1ef0a.slice/crio-6115478fda7d26a508af6d210d0da747a82c4355f9f9d1c51b4fd995382a3278 WatchSource:0}: Error finding container 6115478fda7d26a508af6d210d0da747a82c4355f9f9d1c51b4fd995382a3278: Status 404 returned error can't find the container with id 6115478fda7d26a508af6d210d0da747a82c4355f9f9d1c51b4fd995382a3278 Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.347565 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.349115 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.352160 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.353877 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.354116 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.354246 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-7982z" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.354376 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.388885 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5jgq2"] Dec 15 12:36:32 crc kubenswrapper[4719]: W1215 12:36:32.401438 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5121ea0_ad37_471e_9eca_d88c97375e25.slice/crio-427c9fc89562481b309789460181aaf2ff79fd45c76024b0c6af105489db9ac6 WatchSource:0}: Error finding container 427c9fc89562481b309789460181aaf2ff79fd45c76024b0c6af105489db9ac6: Status 404 returned error can't find the container with id 427c9fc89562481b309789460181aaf2ff79fd45c76024b0c6af105489db9ac6 Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.490895 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b079a1-3e14-4457-aac6-213038651316-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.490945 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b079a1-3e14-4457-aac6-213038651316-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.490984 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h5ln\" (UniqueName: \"kubernetes.io/projected/a0b079a1-3e14-4457-aac6-213038651316-kube-api-access-9h5ln\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.491011 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b079a1-3e14-4457-aac6-213038651316-config\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.491032 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b079a1-3e14-4457-aac6-213038651316-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.491066 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0b079a1-3e14-4457-aac6-213038651316-scripts\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.491095 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a0b079a1-3e14-4457-aac6-213038651316-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.571512 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-00d1-account-create-update-tpf9v"] Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.599738 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0b079a1-3e14-4457-aac6-213038651316-scripts\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.599805 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a0b079a1-3e14-4457-aac6-213038651316-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.599943 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b079a1-3e14-4457-aac6-213038651316-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.599973 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b079a1-3e14-4457-aac6-213038651316-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.600008 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h5ln\" (UniqueName: \"kubernetes.io/projected/a0b079a1-3e14-4457-aac6-213038651316-kube-api-access-9h5ln\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.600033 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b079a1-3e14-4457-aac6-213038651316-config\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.600055 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b079a1-3e14-4457-aac6-213038651316-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.603307 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b079a1-3e14-4457-aac6-213038651316-config\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.603583 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0b079a1-3e14-4457-aac6-213038651316-scripts\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.603815 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a0b079a1-3e14-4457-aac6-213038651316-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.616590 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b079a1-3e14-4457-aac6-213038651316-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.623750 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b079a1-3e14-4457-aac6-213038651316-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.632166 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b079a1-3e14-4457-aac6-213038651316-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.643343 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h5ln\" (UniqueName: \"kubernetes.io/projected/a0b079a1-3e14-4457-aac6-213038651316-kube-api-access-9h5ln\") pod \"ovn-northd-0\" (UID: \"a0b079a1-3e14-4457-aac6-213038651316\") " pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.686954 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.984235 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5jgq2" event={"ID":"e5121ea0-ad37-471e-9eca-d88c97375e25","Type":"ContainerStarted","Data":"e9d54eb7401cd1368de91a626037d8fcc0ae6d7646d63d3a8c03402a94969a37"} Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.984292 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5jgq2" event={"ID":"e5121ea0-ad37-471e-9eca-d88c97375e25","Type":"ContainerStarted","Data":"427c9fc89562481b309789460181aaf2ff79fd45c76024b0c6af105489db9ac6"} Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.987535 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2498-account-create-update-bkr7v" event={"ID":"392caf51-84a6-43d0-9bae-edbf23c1ef0a","Type":"ContainerStarted","Data":"e5cc5f5f146af9ea8b7ade3b01b578bb9e1e5f2762bf9bfb445a5a001521bd04"} Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.987566 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2498-account-create-update-bkr7v" event={"ID":"392caf51-84a6-43d0-9bae-edbf23c1ef0a","Type":"ContainerStarted","Data":"6115478fda7d26a508af6d210d0da747a82c4355f9f9d1c51b4fd995382a3278"} Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.990687 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9mxg6" event={"ID":"709db400-dea1-4f84-a1bd-e3c6bdc44644","Type":"ContainerStarted","Data":"b8e610d478d574b84ff4a581bf6527c98612b4555494e76fce65dfac02699355"} Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.990734 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9mxg6" event={"ID":"709db400-dea1-4f84-a1bd-e3c6bdc44644","Type":"ContainerStarted","Data":"1fcacd54484122528629dbddd6cf20625cca0e10469e9017b9d288db7a0acce3"} Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.992115 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-00d1-account-create-update-tpf9v" event={"ID":"50261891-12d4-4ec9-a0c0-e3b1d07859b6","Type":"ContainerStarted","Data":"f8763d2f1ab31a94b363b104c80cc7d11a71d7999d4b244c60be71c819c74010"} Dec 15 12:36:32 crc kubenswrapper[4719]: I1215 12:36:32.992149 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-00d1-account-create-update-tpf9v" event={"ID":"50261891-12d4-4ec9-a0c0-e3b1d07859b6","Type":"ContainerStarted","Data":"1a7d87a7c8ca38d957ef35da72dc6c56e117f4f0241aefbced4681ceeb71f497"} Dec 15 12:36:33 crc kubenswrapper[4719]: I1215 12:36:33.013604 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-5jgq2" podStartSLOduration=2.013576189 podStartE2EDuration="2.013576189s" podCreationTimestamp="2025-12-15 12:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:36:33.012162584 +0000 UTC m=+1153.954455624" watchObservedRunningTime="2025-12-15 12:36:33.013576189 +0000 UTC m=+1153.955869229" Dec 15 12:36:33 crc kubenswrapper[4719]: I1215 12:36:33.027770 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-9mxg6" podStartSLOduration=2.027748671 podStartE2EDuration="2.027748671s" podCreationTimestamp="2025-12-15 12:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:36:33.025053197 +0000 UTC m=+1153.967346227" watchObservedRunningTime="2025-12-15 12:36:33.027748671 +0000 UTC m=+1153.970041711" Dec 15 12:36:33 crc kubenswrapper[4719]: I1215 12:36:33.048707 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-00d1-account-create-update-tpf9v" podStartSLOduration=2.048691046 podStartE2EDuration="2.048691046s" podCreationTimestamp="2025-12-15 12:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:36:33.044077452 +0000 UTC m=+1153.986370492" watchObservedRunningTime="2025-12-15 12:36:33.048691046 +0000 UTC m=+1153.990984076" Dec 15 12:36:33 crc kubenswrapper[4719]: I1215 12:36:33.057079 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-2498-account-create-update-bkr7v" podStartSLOduration=2.057060598 podStartE2EDuration="2.057060598s" podCreationTimestamp="2025-12-15 12:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:36:33.056057636 +0000 UTC m=+1153.998350666" watchObservedRunningTime="2025-12-15 12:36:33.057060598 +0000 UTC m=+1153.999353638" Dec 15 12:36:33 crc kubenswrapper[4719]: I1215 12:36:33.182078 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 15 12:36:33 crc kubenswrapper[4719]: I1215 12:36:33.383149 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 15 12:36:34 crc kubenswrapper[4719]: I1215 12:36:34.011322 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a0b079a1-3e14-4457-aac6-213038651316","Type":"ContainerStarted","Data":"18096a67c0686b3a6a64304f24f11a0eedd23a8427ce430eaf2b99943ef01eb5"} Dec 15 12:36:34 crc kubenswrapper[4719]: I1215 12:36:34.017000 4719 generic.go:334] "Generic (PLEG): container finished" podID="392caf51-84a6-43d0-9bae-edbf23c1ef0a" containerID="e5cc5f5f146af9ea8b7ade3b01b578bb9e1e5f2762bf9bfb445a5a001521bd04" exitCode=0 Dec 15 12:36:34 crc kubenswrapper[4719]: I1215 12:36:34.017096 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2498-account-create-update-bkr7v" event={"ID":"392caf51-84a6-43d0-9bae-edbf23c1ef0a","Type":"ContainerDied","Data":"e5cc5f5f146af9ea8b7ade3b01b578bb9e1e5f2762bf9bfb445a5a001521bd04"} Dec 15 12:36:34 crc kubenswrapper[4719]: I1215 12:36:34.018986 4719 generic.go:334] "Generic (PLEG): container finished" podID="709db400-dea1-4f84-a1bd-e3c6bdc44644" containerID="b8e610d478d574b84ff4a581bf6527c98612b4555494e76fce65dfac02699355" exitCode=0 Dec 15 12:36:34 crc kubenswrapper[4719]: I1215 12:36:34.019075 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9mxg6" event={"ID":"709db400-dea1-4f84-a1bd-e3c6bdc44644","Type":"ContainerDied","Data":"b8e610d478d574b84ff4a581bf6527c98612b4555494e76fce65dfac02699355"} Dec 15 12:36:34 crc kubenswrapper[4719]: I1215 12:36:34.020420 4719 generic.go:334] "Generic (PLEG): container finished" podID="50261891-12d4-4ec9-a0c0-e3b1d07859b6" containerID="f8763d2f1ab31a94b363b104c80cc7d11a71d7999d4b244c60be71c819c74010" exitCode=0 Dec 15 12:36:34 crc kubenswrapper[4719]: I1215 12:36:34.020492 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-00d1-account-create-update-tpf9v" event={"ID":"50261891-12d4-4ec9-a0c0-e3b1d07859b6","Type":"ContainerDied","Data":"f8763d2f1ab31a94b363b104c80cc7d11a71d7999d4b244c60be71c819c74010"} Dec 15 12:36:34 crc kubenswrapper[4719]: I1215 12:36:34.027420 4719 generic.go:334] "Generic (PLEG): container finished" podID="e5121ea0-ad37-471e-9eca-d88c97375e25" containerID="e9d54eb7401cd1368de91a626037d8fcc0ae6d7646d63d3a8c03402a94969a37" exitCode=0 Dec 15 12:36:34 crc kubenswrapper[4719]: I1215 12:36:34.027507 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5jgq2" event={"ID":"e5121ea0-ad37-471e-9eca-d88c97375e25","Type":"ContainerDied","Data":"e9d54eb7401cd1368de91a626037d8fcc0ae6d7646d63d3a8c03402a94969a37"} Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.034966 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a0b079a1-3e14-4457-aac6-213038651316","Type":"ContainerStarted","Data":"f9a9670d76d2c619e4fc7fe865ae63fb29c3a7c6b61eecf53f186a43d84bbfdf"} Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.035308 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a0b079a1-3e14-4457-aac6-213038651316","Type":"ContainerStarted","Data":"990f07c1b9504f296f2158a83f6db4e08e02ebade5b9f40e09d289d1b737bd7d"} Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.072220 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.642221476 podStartE2EDuration="3.072193799s" podCreationTimestamp="2025-12-15 12:36:32 +0000 UTC" firstStartedPulling="2025-12-15 12:36:33.228592992 +0000 UTC m=+1154.170886022" lastFinishedPulling="2025-12-15 12:36:34.658565315 +0000 UTC m=+1155.600858345" observedRunningTime="2025-12-15 12:36:35.064782097 +0000 UTC m=+1156.007075127" watchObservedRunningTime="2025-12-15 12:36:35.072193799 +0000 UTC m=+1156.014486829" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.430555 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5jgq2" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.567980 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrzwd\" (UniqueName: \"kubernetes.io/projected/e5121ea0-ad37-471e-9eca-d88c97375e25-kube-api-access-mrzwd\") pod \"e5121ea0-ad37-471e-9eca-d88c97375e25\" (UID: \"e5121ea0-ad37-471e-9eca-d88c97375e25\") " Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.568260 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5121ea0-ad37-471e-9eca-d88c97375e25-operator-scripts\") pod \"e5121ea0-ad37-471e-9eca-d88c97375e25\" (UID: \"e5121ea0-ad37-471e-9eca-d88c97375e25\") " Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.569642 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-00d1-account-create-update-tpf9v" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.569949 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5121ea0-ad37-471e-9eca-d88c97375e25-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5121ea0-ad37-471e-9eca-d88c97375e25" (UID: "e5121ea0-ad37-471e-9eca-d88c97375e25"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.574868 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5121ea0-ad37-471e-9eca-d88c97375e25-kube-api-access-mrzwd" (OuterVolumeSpecName: "kube-api-access-mrzwd") pod "e5121ea0-ad37-471e-9eca-d88c97375e25" (UID: "e5121ea0-ad37-471e-9eca-d88c97375e25"). InnerVolumeSpecName "kube-api-access-mrzwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.600227 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9mxg6" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.641117 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2498-account-create-update-bkr7v" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.669623 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/709db400-dea1-4f84-a1bd-e3c6bdc44644-operator-scripts\") pod \"709db400-dea1-4f84-a1bd-e3c6bdc44644\" (UID: \"709db400-dea1-4f84-a1bd-e3c6bdc44644\") " Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.669765 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50261891-12d4-4ec9-a0c0-e3b1d07859b6-operator-scripts\") pod \"50261891-12d4-4ec9-a0c0-e3b1d07859b6\" (UID: \"50261891-12d4-4ec9-a0c0-e3b1d07859b6\") " Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.670235 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50261891-12d4-4ec9-a0c0-e3b1d07859b6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "50261891-12d4-4ec9-a0c0-e3b1d07859b6" (UID: "50261891-12d4-4ec9-a0c0-e3b1d07859b6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.670350 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpkcl\" (UniqueName: \"kubernetes.io/projected/709db400-dea1-4f84-a1bd-e3c6bdc44644-kube-api-access-qpkcl\") pod \"709db400-dea1-4f84-a1bd-e3c6bdc44644\" (UID: \"709db400-dea1-4f84-a1bd-e3c6bdc44644\") " Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.670739 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxr5h\" (UniqueName: \"kubernetes.io/projected/50261891-12d4-4ec9-a0c0-e3b1d07859b6-kube-api-access-sxr5h\") pod \"50261891-12d4-4ec9-a0c0-e3b1d07859b6\" (UID: \"50261891-12d4-4ec9-a0c0-e3b1d07859b6\") " Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.670975 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/709db400-dea1-4f84-a1bd-e3c6bdc44644-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "709db400-dea1-4f84-a1bd-e3c6bdc44644" (UID: "709db400-dea1-4f84-a1bd-e3c6bdc44644"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.671226 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5121ea0-ad37-471e-9eca-d88c97375e25-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.671239 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/709db400-dea1-4f84-a1bd-e3c6bdc44644-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.671249 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrzwd\" (UniqueName: \"kubernetes.io/projected/e5121ea0-ad37-471e-9eca-d88c97375e25-kube-api-access-mrzwd\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.671258 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50261891-12d4-4ec9-a0c0-e3b1d07859b6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.673655 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/709db400-dea1-4f84-a1bd-e3c6bdc44644-kube-api-access-qpkcl" (OuterVolumeSpecName: "kube-api-access-qpkcl") pod "709db400-dea1-4f84-a1bd-e3c6bdc44644" (UID: "709db400-dea1-4f84-a1bd-e3c6bdc44644"). InnerVolumeSpecName "kube-api-access-qpkcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.682127 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50261891-12d4-4ec9-a0c0-e3b1d07859b6-kube-api-access-sxr5h" (OuterVolumeSpecName: "kube-api-access-sxr5h") pod "50261891-12d4-4ec9-a0c0-e3b1d07859b6" (UID: "50261891-12d4-4ec9-a0c0-e3b1d07859b6"). InnerVolumeSpecName "kube-api-access-sxr5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.772176 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/392caf51-84a6-43d0-9bae-edbf23c1ef0a-operator-scripts\") pod \"392caf51-84a6-43d0-9bae-edbf23c1ef0a\" (UID: \"392caf51-84a6-43d0-9bae-edbf23c1ef0a\") " Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.772232 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8ghm\" (UniqueName: \"kubernetes.io/projected/392caf51-84a6-43d0-9bae-edbf23c1ef0a-kube-api-access-t8ghm\") pod \"392caf51-84a6-43d0-9bae-edbf23c1ef0a\" (UID: \"392caf51-84a6-43d0-9bae-edbf23c1ef0a\") " Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.772582 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpkcl\" (UniqueName: \"kubernetes.io/projected/709db400-dea1-4f84-a1bd-e3c6bdc44644-kube-api-access-qpkcl\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.772605 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxr5h\" (UniqueName: \"kubernetes.io/projected/50261891-12d4-4ec9-a0c0-e3b1d07859b6-kube-api-access-sxr5h\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.772836 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/392caf51-84a6-43d0-9bae-edbf23c1ef0a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "392caf51-84a6-43d0-9bae-edbf23c1ef0a" (UID: "392caf51-84a6-43d0-9bae-edbf23c1ef0a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.775398 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/392caf51-84a6-43d0-9bae-edbf23c1ef0a-kube-api-access-t8ghm" (OuterVolumeSpecName: "kube-api-access-t8ghm") pod "392caf51-84a6-43d0-9bae-edbf23c1ef0a" (UID: "392caf51-84a6-43d0-9bae-edbf23c1ef0a"). InnerVolumeSpecName "kube-api-access-t8ghm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.873829 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/392caf51-84a6-43d0-9bae-edbf23c1ef0a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:35 crc kubenswrapper[4719]: I1215 12:36:35.873894 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8ghm\" (UniqueName: \"kubernetes.io/projected/392caf51-84a6-43d0-9bae-edbf23c1ef0a-kube-api-access-t8ghm\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.045096 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-00d1-account-create-update-tpf9v" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.045497 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-00d1-account-create-update-tpf9v" event={"ID":"50261891-12d4-4ec9-a0c0-e3b1d07859b6","Type":"ContainerDied","Data":"1a7d87a7c8ca38d957ef35da72dc6c56e117f4f0241aefbced4681ceeb71f497"} Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.045534 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a7d87a7c8ca38d957ef35da72dc6c56e117f4f0241aefbced4681ceeb71f497" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.055879 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5jgq2" event={"ID":"e5121ea0-ad37-471e-9eca-d88c97375e25","Type":"ContainerDied","Data":"427c9fc89562481b309789460181aaf2ff79fd45c76024b0c6af105489db9ac6"} Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.055914 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="427c9fc89562481b309789460181aaf2ff79fd45c76024b0c6af105489db9ac6" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.056509 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5jgq2" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.058833 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2498-account-create-update-bkr7v" event={"ID":"392caf51-84a6-43d0-9bae-edbf23c1ef0a","Type":"ContainerDied","Data":"6115478fda7d26a508af6d210d0da747a82c4355f9f9d1c51b4fd995382a3278"} Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.058886 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6115478fda7d26a508af6d210d0da747a82c4355f9f9d1c51b4fd995382a3278" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.058842 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2498-account-create-update-bkr7v" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.060393 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9mxg6" event={"ID":"709db400-dea1-4f84-a1bd-e3c6bdc44644","Type":"ContainerDied","Data":"1fcacd54484122528629dbddd6cf20625cca0e10469e9017b9d288db7a0acce3"} Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.060421 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fcacd54484122528629dbddd6cf20625cca0e10469e9017b9d288db7a0acce3" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.060407 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9mxg6" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.060738 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.529176 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.644004 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.830991 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:36 crc kubenswrapper[4719]: I1215 12:36:36.954334 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f6tsl"] Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.066545 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" podUID="15c6924c-50fe-418b-bbde-f890493f82ed" containerName="dnsmasq-dns" containerID="cri-o://d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030" gracePeriod=10 Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.068278 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-dndhn"] Dec 15 12:36:37 crc kubenswrapper[4719]: E1215 12:36:37.069366 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5121ea0-ad37-471e-9eca-d88c97375e25" containerName="mariadb-database-create" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.069439 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5121ea0-ad37-471e-9eca-d88c97375e25" containerName="mariadb-database-create" Dec 15 12:36:37 crc kubenswrapper[4719]: E1215 12:36:37.069514 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50261891-12d4-4ec9-a0c0-e3b1d07859b6" containerName="mariadb-account-create-update" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.069582 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="50261891-12d4-4ec9-a0c0-e3b1d07859b6" containerName="mariadb-account-create-update" Dec 15 12:36:37 crc kubenswrapper[4719]: E1215 12:36:37.069646 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709db400-dea1-4f84-a1bd-e3c6bdc44644" containerName="mariadb-database-create" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.069705 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="709db400-dea1-4f84-a1bd-e3c6bdc44644" containerName="mariadb-database-create" Dec 15 12:36:37 crc kubenswrapper[4719]: E1215 12:36:37.069778 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392caf51-84a6-43d0-9bae-edbf23c1ef0a" containerName="mariadb-account-create-update" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.069831 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="392caf51-84a6-43d0-9bae-edbf23c1ef0a" containerName="mariadb-account-create-update" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.070044 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5121ea0-ad37-471e-9eca-d88c97375e25" containerName="mariadb-database-create" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.070192 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="709db400-dea1-4f84-a1bd-e3c6bdc44644" containerName="mariadb-database-create" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.073145 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="392caf51-84a6-43d0-9bae-edbf23c1ef0a" containerName="mariadb-account-create-update" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.073286 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="50261891-12d4-4ec9-a0c0-e3b1d07859b6" containerName="mariadb-account-create-update" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.073966 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dndhn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.100465 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dndhn"] Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.168782 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-2228-account-create-update-rwhbn"] Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.170058 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2228-account-create-update-rwhbn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.175905 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.189000 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2228-account-create-update-rwhbn"] Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.195957 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxrm4\" (UniqueName: \"kubernetes.io/projected/eaaec482-1c98-40c9-a402-97e6b8241a98-kube-api-access-pxrm4\") pod \"glance-db-create-dndhn\" (UID: \"eaaec482-1c98-40c9-a402-97e6b8241a98\") " pod="openstack/glance-db-create-dndhn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.196047 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaaec482-1c98-40c9-a402-97e6b8241a98-operator-scripts\") pod \"glance-db-create-dndhn\" (UID: \"eaaec482-1c98-40c9-a402-97e6b8241a98\") " pod="openstack/glance-db-create-dndhn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.297615 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxrm4\" (UniqueName: \"kubernetes.io/projected/eaaec482-1c98-40c9-a402-97e6b8241a98-kube-api-access-pxrm4\") pod \"glance-db-create-dndhn\" (UID: \"eaaec482-1c98-40c9-a402-97e6b8241a98\") " pod="openstack/glance-db-create-dndhn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.297683 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpbgf\" (UniqueName: \"kubernetes.io/projected/7238f625-5678-487a-91fc-a3a7686441bd-kube-api-access-kpbgf\") pod \"glance-2228-account-create-update-rwhbn\" (UID: \"7238f625-5678-487a-91fc-a3a7686441bd\") " pod="openstack/glance-2228-account-create-update-rwhbn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.297718 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaaec482-1c98-40c9-a402-97e6b8241a98-operator-scripts\") pod \"glance-db-create-dndhn\" (UID: \"eaaec482-1c98-40c9-a402-97e6b8241a98\") " pod="openstack/glance-db-create-dndhn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.297749 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7238f625-5678-487a-91fc-a3a7686441bd-operator-scripts\") pod \"glance-2228-account-create-update-rwhbn\" (UID: \"7238f625-5678-487a-91fc-a3a7686441bd\") " pod="openstack/glance-2228-account-create-update-rwhbn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.298899 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaaec482-1c98-40c9-a402-97e6b8241a98-operator-scripts\") pod \"glance-db-create-dndhn\" (UID: \"eaaec482-1c98-40c9-a402-97e6b8241a98\") " pod="openstack/glance-db-create-dndhn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.320163 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxrm4\" (UniqueName: \"kubernetes.io/projected/eaaec482-1c98-40c9-a402-97e6b8241a98-kube-api-access-pxrm4\") pod \"glance-db-create-dndhn\" (UID: \"eaaec482-1c98-40c9-a402-97e6b8241a98\") " pod="openstack/glance-db-create-dndhn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.398656 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7238f625-5678-487a-91fc-a3a7686441bd-operator-scripts\") pod \"glance-2228-account-create-update-rwhbn\" (UID: \"7238f625-5678-487a-91fc-a3a7686441bd\") " pod="openstack/glance-2228-account-create-update-rwhbn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.398787 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpbgf\" (UniqueName: \"kubernetes.io/projected/7238f625-5678-487a-91fc-a3a7686441bd-kube-api-access-kpbgf\") pod \"glance-2228-account-create-update-rwhbn\" (UID: \"7238f625-5678-487a-91fc-a3a7686441bd\") " pod="openstack/glance-2228-account-create-update-rwhbn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.399229 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dndhn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.399636 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7238f625-5678-487a-91fc-a3a7686441bd-operator-scripts\") pod \"glance-2228-account-create-update-rwhbn\" (UID: \"7238f625-5678-487a-91fc-a3a7686441bd\") " pod="openstack/glance-2228-account-create-update-rwhbn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.428396 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpbgf\" (UniqueName: \"kubernetes.io/projected/7238f625-5678-487a-91fc-a3a7686441bd-kube-api-access-kpbgf\") pod \"glance-2228-account-create-update-rwhbn\" (UID: \"7238f625-5678-487a-91fc-a3a7686441bd\") " pod="openstack/glance-2228-account-create-update-rwhbn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.545034 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2228-account-create-update-rwhbn" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.565228 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.703924 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcmms\" (UniqueName: \"kubernetes.io/projected/15c6924c-50fe-418b-bbde-f890493f82ed-kube-api-access-gcmms\") pod \"15c6924c-50fe-418b-bbde-f890493f82ed\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.704075 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-ovsdbserver-nb\") pod \"15c6924c-50fe-418b-bbde-f890493f82ed\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.704101 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-config\") pod \"15c6924c-50fe-418b-bbde-f890493f82ed\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.704368 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-dns-svc\") pod \"15c6924c-50fe-418b-bbde-f890493f82ed\" (UID: \"15c6924c-50fe-418b-bbde-f890493f82ed\") " Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.711445 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15c6924c-50fe-418b-bbde-f890493f82ed-kube-api-access-gcmms" (OuterVolumeSpecName: "kube-api-access-gcmms") pod "15c6924c-50fe-418b-bbde-f890493f82ed" (UID: "15c6924c-50fe-418b-bbde-f890493f82ed"). InnerVolumeSpecName "kube-api-access-gcmms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.761110 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-config" (OuterVolumeSpecName: "config") pod "15c6924c-50fe-418b-bbde-f890493f82ed" (UID: "15c6924c-50fe-418b-bbde-f890493f82ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.761254 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "15c6924c-50fe-418b-bbde-f890493f82ed" (UID: "15c6924c-50fe-418b-bbde-f890493f82ed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.761339 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "15c6924c-50fe-418b-bbde-f890493f82ed" (UID: "15c6924c-50fe-418b-bbde-f890493f82ed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.807762 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcmms\" (UniqueName: \"kubernetes.io/projected/15c6924c-50fe-418b-bbde-f890493f82ed-kube-api-access-gcmms\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.807888 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.807902 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.807920 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15c6924c-50fe-418b-bbde-f890493f82ed-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:37 crc kubenswrapper[4719]: I1215 12:36:37.971826 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dndhn"] Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.088107 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dndhn" event={"ID":"eaaec482-1c98-40c9-a402-97e6b8241a98","Type":"ContainerStarted","Data":"2ddce473473f971aafe9b9e44907e83316d64c60b2d24c596749c80a6abe4ce3"} Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.092254 4719 generic.go:334] "Generic (PLEG): container finished" podID="15c6924c-50fe-418b-bbde-f890493f82ed" containerID="d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030" exitCode=0 Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.092302 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" event={"ID":"15c6924c-50fe-418b-bbde-f890493f82ed","Type":"ContainerDied","Data":"d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030"} Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.092331 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" event={"ID":"15c6924c-50fe-418b-bbde-f890493f82ed","Type":"ContainerDied","Data":"22c66f14ccca7c4873e9a034cb36a314adcb8f7385df742fc58548892dc2fc31"} Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.092350 4719 scope.go:117] "RemoveContainer" containerID="d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030" Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.092499 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f6tsl" Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.099480 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2228-account-create-update-rwhbn"] Dec 15 12:36:38 crc kubenswrapper[4719]: W1215 12:36:38.102674 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7238f625_5678_487a_91fc_a3a7686441bd.slice/crio-9717c3a1f3964f24778babd69a66cf1daaf280d2342e941c35c24694b4c74692 WatchSource:0}: Error finding container 9717c3a1f3964f24778babd69a66cf1daaf280d2342e941c35c24694b4c74692: Status 404 returned error can't find the container with id 9717c3a1f3964f24778babd69a66cf1daaf280d2342e941c35c24694b4c74692 Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.114313 4719 scope.go:117] "RemoveContainer" containerID="9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e" Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.130689 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f6tsl"] Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.135627 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f6tsl"] Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.148060 4719 scope.go:117] "RemoveContainer" containerID="d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030" Dec 15 12:36:38 crc kubenswrapper[4719]: E1215 12:36:38.148572 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030\": container with ID starting with d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030 not found: ID does not exist" containerID="d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030" Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.148612 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030"} err="failed to get container status \"d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030\": rpc error: code = NotFound desc = could not find container \"d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030\": container with ID starting with d61ed8660f80b1537253c79e5ad1f9b1611a6f4521332c3ca9010576ab7bd030 not found: ID does not exist" Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.148638 4719 scope.go:117] "RemoveContainer" containerID="9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e" Dec 15 12:36:38 crc kubenswrapper[4719]: E1215 12:36:38.148912 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e\": container with ID starting with 9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e not found: ID does not exist" containerID="9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e" Dec 15 12:36:38 crc kubenswrapper[4719]: I1215 12:36:38.148945 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e"} err="failed to get container status \"9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e\": rpc error: code = NotFound desc = could not find container \"9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e\": container with ID starting with 9795f9ce86ce5352816e8a4d6c306f552557778ccdc30ce20d807f99277cea1e not found: ID does not exist" Dec 15 12:36:39 crc kubenswrapper[4719]: I1215 12:36:39.103045 4719 generic.go:334] "Generic (PLEG): container finished" podID="7238f625-5678-487a-91fc-a3a7686441bd" containerID="96c75bdc1d8541db7b6e7a92f093ee49a969e8d1d327ae74d40290660ce7c297" exitCode=0 Dec 15 12:36:39 crc kubenswrapper[4719]: I1215 12:36:39.103111 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2228-account-create-update-rwhbn" event={"ID":"7238f625-5678-487a-91fc-a3a7686441bd","Type":"ContainerDied","Data":"96c75bdc1d8541db7b6e7a92f093ee49a969e8d1d327ae74d40290660ce7c297"} Dec 15 12:36:39 crc kubenswrapper[4719]: I1215 12:36:39.103453 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2228-account-create-update-rwhbn" event={"ID":"7238f625-5678-487a-91fc-a3a7686441bd","Type":"ContainerStarted","Data":"9717c3a1f3964f24778babd69a66cf1daaf280d2342e941c35c24694b4c74692"} Dec 15 12:36:39 crc kubenswrapper[4719]: I1215 12:36:39.105433 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaaec482-1c98-40c9-a402-97e6b8241a98" containerID="6e1db7d818f96e2c02ed751e956127418fd37244d77f05c42873cad1ce04282d" exitCode=0 Dec 15 12:36:39 crc kubenswrapper[4719]: I1215 12:36:39.105469 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dndhn" event={"ID":"eaaec482-1c98-40c9-a402-97e6b8241a98","Type":"ContainerDied","Data":"6e1db7d818f96e2c02ed751e956127418fd37244d77f05c42873cad1ce04282d"} Dec 15 12:36:39 crc kubenswrapper[4719]: I1215 12:36:39.485617 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15c6924c-50fe-418b-bbde-f890493f82ed" path="/var/lib/kubelet/pods/15c6924c-50fe-418b-bbde-f890493f82ed/volumes" Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.518895 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dndhn" Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.531713 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2228-account-create-update-rwhbn" Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.652602 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpbgf\" (UniqueName: \"kubernetes.io/projected/7238f625-5678-487a-91fc-a3a7686441bd-kube-api-access-kpbgf\") pod \"7238f625-5678-487a-91fc-a3a7686441bd\" (UID: \"7238f625-5678-487a-91fc-a3a7686441bd\") " Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.652707 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaaec482-1c98-40c9-a402-97e6b8241a98-operator-scripts\") pod \"eaaec482-1c98-40c9-a402-97e6b8241a98\" (UID: \"eaaec482-1c98-40c9-a402-97e6b8241a98\") " Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.652752 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7238f625-5678-487a-91fc-a3a7686441bd-operator-scripts\") pod \"7238f625-5678-487a-91fc-a3a7686441bd\" (UID: \"7238f625-5678-487a-91fc-a3a7686441bd\") " Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.652913 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxrm4\" (UniqueName: \"kubernetes.io/projected/eaaec482-1c98-40c9-a402-97e6b8241a98-kube-api-access-pxrm4\") pod \"eaaec482-1c98-40c9-a402-97e6b8241a98\" (UID: \"eaaec482-1c98-40c9-a402-97e6b8241a98\") " Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.654139 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7238f625-5678-487a-91fc-a3a7686441bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7238f625-5678-487a-91fc-a3a7686441bd" (UID: "7238f625-5678-487a-91fc-a3a7686441bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.654210 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eaaec482-1c98-40c9-a402-97e6b8241a98-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eaaec482-1c98-40c9-a402-97e6b8241a98" (UID: "eaaec482-1c98-40c9-a402-97e6b8241a98"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.660072 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7238f625-5678-487a-91fc-a3a7686441bd-kube-api-access-kpbgf" (OuterVolumeSpecName: "kube-api-access-kpbgf") pod "7238f625-5678-487a-91fc-a3a7686441bd" (UID: "7238f625-5678-487a-91fc-a3a7686441bd"). InnerVolumeSpecName "kube-api-access-kpbgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.660156 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaaec482-1c98-40c9-a402-97e6b8241a98-kube-api-access-pxrm4" (OuterVolumeSpecName: "kube-api-access-pxrm4") pod "eaaec482-1c98-40c9-a402-97e6b8241a98" (UID: "eaaec482-1c98-40c9-a402-97e6b8241a98"). InnerVolumeSpecName "kube-api-access-pxrm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.754199 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxrm4\" (UniqueName: \"kubernetes.io/projected/eaaec482-1c98-40c9-a402-97e6b8241a98-kube-api-access-pxrm4\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.754412 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpbgf\" (UniqueName: \"kubernetes.io/projected/7238f625-5678-487a-91fc-a3a7686441bd-kube-api-access-kpbgf\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.754469 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eaaec482-1c98-40c9-a402-97e6b8241a98-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:40 crc kubenswrapper[4719]: I1215 12:36:40.754527 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7238f625-5678-487a-91fc-a3a7686441bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:41 crc kubenswrapper[4719]: I1215 12:36:41.128052 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2228-account-create-update-rwhbn" event={"ID":"7238f625-5678-487a-91fc-a3a7686441bd","Type":"ContainerDied","Data":"9717c3a1f3964f24778babd69a66cf1daaf280d2342e941c35c24694b4c74692"} Dec 15 12:36:41 crc kubenswrapper[4719]: I1215 12:36:41.128095 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9717c3a1f3964f24778babd69a66cf1daaf280d2342e941c35c24694b4c74692" Dec 15 12:36:41 crc kubenswrapper[4719]: I1215 12:36:41.128164 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2228-account-create-update-rwhbn" Dec 15 12:36:41 crc kubenswrapper[4719]: I1215 12:36:41.129977 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dndhn" event={"ID":"eaaec482-1c98-40c9-a402-97e6b8241a98","Type":"ContainerDied","Data":"2ddce473473f971aafe9b9e44907e83316d64c60b2d24c596749c80a6abe4ce3"} Dec 15 12:36:41 crc kubenswrapper[4719]: I1215 12:36:41.130071 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ddce473473f971aafe9b9e44907e83316d64c60b2d24c596749c80a6abe4ce3" Dec 15 12:36:41 crc kubenswrapper[4719]: I1215 12:36:41.130219 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dndhn" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.329645 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-nf7qt"] Dec 15 12:36:42 crc kubenswrapper[4719]: E1215 12:36:42.330376 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c6924c-50fe-418b-bbde-f890493f82ed" containerName="dnsmasq-dns" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.330393 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c6924c-50fe-418b-bbde-f890493f82ed" containerName="dnsmasq-dns" Dec 15 12:36:42 crc kubenswrapper[4719]: E1215 12:36:42.330422 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaaec482-1c98-40c9-a402-97e6b8241a98" containerName="mariadb-database-create" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.330431 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaaec482-1c98-40c9-a402-97e6b8241a98" containerName="mariadb-database-create" Dec 15 12:36:42 crc kubenswrapper[4719]: E1215 12:36:42.330456 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c6924c-50fe-418b-bbde-f890493f82ed" containerName="init" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.330465 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c6924c-50fe-418b-bbde-f890493f82ed" containerName="init" Dec 15 12:36:42 crc kubenswrapper[4719]: E1215 12:36:42.330480 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7238f625-5678-487a-91fc-a3a7686441bd" containerName="mariadb-account-create-update" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.330488 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="7238f625-5678-487a-91fc-a3a7686441bd" containerName="mariadb-account-create-update" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.330660 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="7238f625-5678-487a-91fc-a3a7686441bd" containerName="mariadb-account-create-update" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.330677 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c6924c-50fe-418b-bbde-f890493f82ed" containerName="dnsmasq-dns" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.330694 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaaec482-1c98-40c9-a402-97e6b8241a98" containerName="mariadb-database-create" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.331342 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.333521 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.344849 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-nf7qt"] Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.346286 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-npxlb" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.481983 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-db-sync-config-data\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.482147 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-config-data\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.482184 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgqq8\" (UniqueName: \"kubernetes.io/projected/479f222a-bfc2-4257-9588-941538d75171-kube-api-access-jgqq8\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.482221 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-combined-ca-bundle\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.583225 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-config-data\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.583268 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgqq8\" (UniqueName: \"kubernetes.io/projected/479f222a-bfc2-4257-9588-941538d75171-kube-api-access-jgqq8\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.583301 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-combined-ca-bundle\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.583338 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-db-sync-config-data\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.590681 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-db-sync-config-data\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.590765 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-combined-ca-bundle\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.591226 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-config-data\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.608034 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgqq8\" (UniqueName: \"kubernetes.io/projected/479f222a-bfc2-4257-9588-941538d75171-kube-api-access-jgqq8\") pod \"glance-db-sync-nf7qt\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:42 crc kubenswrapper[4719]: I1215 12:36:42.647416 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nf7qt" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.154592 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-nf7qt"] Dec 15 12:36:43 crc kubenswrapper[4719]: W1215 12:36:43.157122 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod479f222a_bfc2_4257_9588_941538d75171.slice/crio-5479d7b82227bac705b45f45fae85af34a996ecafcdcd3f96acab69b64bb1b58 WatchSource:0}: Error finding container 5479d7b82227bac705b45f45fae85af34a996ecafcdcd3f96acab69b64bb1b58: Status 404 returned error can't find the container with id 5479d7b82227bac705b45f45fae85af34a996ecafcdcd3f96acab69b64bb1b58 Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.473244 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-g4d2t"] Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.475176 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.499369 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-g4d2t"] Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.599411 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.599607 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-dns-svc\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.599771 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-config\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.599903 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9vw4\" (UniqueName: \"kubernetes.io/projected/dda77ba2-3b59-4946-940b-8c8987f2884b-kube-api-access-b9vw4\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.600152 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.701372 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9vw4\" (UniqueName: \"kubernetes.io/projected/dda77ba2-3b59-4946-940b-8c8987f2884b-kube-api-access-b9vw4\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.701471 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.701540 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.701591 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-dns-svc\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.701636 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-config\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.703258 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-dns-svc\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.703305 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.703374 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-config\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.703739 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.751009 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9vw4\" (UniqueName: \"kubernetes.io/projected/dda77ba2-3b59-4946-940b-8c8987f2884b-kube-api-access-b9vw4\") pod \"dnsmasq-dns-698758b865-g4d2t\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:43 crc kubenswrapper[4719]: I1215 12:36:43.796265 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.156622 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nf7qt" event={"ID":"479f222a-bfc2-4257-9588-941538d75171","Type":"ContainerStarted","Data":"5479d7b82227bac705b45f45fae85af34a996ecafcdcd3f96acab69b64bb1b58"} Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.393849 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-g4d2t"] Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.695621 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.700827 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.705500 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-hblwj" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.705654 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.709171 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.709314 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.721272 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.823889 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.823929 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5qsb\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-kube-api-access-q5qsb\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.824010 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.824141 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/651edea8-3300-463c-a0fa-abd75f6056e0-lock\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.824188 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/651edea8-3300-463c-a0fa-abd75f6056e0-cache\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: E1215 12:36:44.833492 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddda77ba2_3b59_4946_940b_8c8987f2884b.slice/crio-ba4129088dbebcd0ed5673846e7fe28747d58d295ddfca0b1402ed057144f27e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddda77ba2_3b59_4946_940b_8c8987f2884b.slice/crio-conmon-ba4129088dbebcd0ed5673846e7fe28747d58d295ddfca0b1402ed057144f27e.scope\": RecentStats: unable to find data in memory cache]" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.926083 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/651edea8-3300-463c-a0fa-abd75f6056e0-cache\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.926448 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.926468 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5qsb\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-kube-api-access-q5qsb\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.926535 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/651edea8-3300-463c-a0fa-abd75f6056e0-cache\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: E1215 12:36:44.926609 4719 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 15 12:36:44 crc kubenswrapper[4719]: E1215 12:36:44.926625 4719 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 15 12:36:44 crc kubenswrapper[4719]: E1215 12:36:44.927107 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift podName:651edea8-3300-463c-a0fa-abd75f6056e0 nodeName:}" failed. No retries permitted until 2025-12-15 12:36:45.427081628 +0000 UTC m=+1166.369374648 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift") pod "swift-storage-0" (UID: "651edea8-3300-463c-a0fa-abd75f6056e0") : configmap "swift-ring-files" not found Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.927248 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.934001 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.934319 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/651edea8-3300-463c-a0fa-abd75f6056e0-lock\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.934717 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/651edea8-3300-463c-a0fa-abd75f6056e0-lock\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.951173 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:44 crc kubenswrapper[4719]: I1215 12:36:44.954420 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5qsb\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-kube-api-access-q5qsb\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:45 crc kubenswrapper[4719]: I1215 12:36:45.164593 4719 generic.go:334] "Generic (PLEG): container finished" podID="dda77ba2-3b59-4946-940b-8c8987f2884b" containerID="ba4129088dbebcd0ed5673846e7fe28747d58d295ddfca0b1402ed057144f27e" exitCode=0 Dec 15 12:36:45 crc kubenswrapper[4719]: I1215 12:36:45.164646 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-g4d2t" event={"ID":"dda77ba2-3b59-4946-940b-8c8987f2884b","Type":"ContainerDied","Data":"ba4129088dbebcd0ed5673846e7fe28747d58d295ddfca0b1402ed057144f27e"} Dec 15 12:36:45 crc kubenswrapper[4719]: I1215 12:36:45.164715 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-g4d2t" event={"ID":"dda77ba2-3b59-4946-940b-8c8987f2884b","Type":"ContainerStarted","Data":"379e77f8161bcdb483565f2d1ef2c586af1cda5db00b990b6150d2f18e1906d4"} Dec 15 12:36:45 crc kubenswrapper[4719]: I1215 12:36:45.442046 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:45 crc kubenswrapper[4719]: E1215 12:36:45.442621 4719 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 15 12:36:45 crc kubenswrapper[4719]: E1215 12:36:45.442635 4719 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 15 12:36:45 crc kubenswrapper[4719]: E1215 12:36:45.442679 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift podName:651edea8-3300-463c-a0fa-abd75f6056e0 nodeName:}" failed. No retries permitted until 2025-12-15 12:36:46.4426662 +0000 UTC m=+1167.384959230 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift") pod "swift-storage-0" (UID: "651edea8-3300-463c-a0fa-abd75f6056e0") : configmap "swift-ring-files" not found Dec 15 12:36:46 crc kubenswrapper[4719]: I1215 12:36:46.178105 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-g4d2t" event={"ID":"dda77ba2-3b59-4946-940b-8c8987f2884b","Type":"ContainerStarted","Data":"1e9f71ff98b60da5c1ac92b2eb5821eba7caa3d3a4c4b54b042ea00689040490"} Dec 15 12:36:46 crc kubenswrapper[4719]: I1215 12:36:46.178293 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:46 crc kubenswrapper[4719]: I1215 12:36:46.203611 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-g4d2t" podStartSLOduration=3.203587378 podStartE2EDuration="3.203587378s" podCreationTimestamp="2025-12-15 12:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:36:46.194435543 +0000 UTC m=+1167.136728573" watchObservedRunningTime="2025-12-15 12:36:46.203587378 +0000 UTC m=+1167.145880418" Dec 15 12:36:46 crc kubenswrapper[4719]: I1215 12:36:46.465971 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:46 crc kubenswrapper[4719]: E1215 12:36:46.466128 4719 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 15 12:36:46 crc kubenswrapper[4719]: E1215 12:36:46.466146 4719 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 15 12:36:46 crc kubenswrapper[4719]: E1215 12:36:46.466195 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift podName:651edea8-3300-463c-a0fa-abd75f6056e0 nodeName:}" failed. No retries permitted until 2025-12-15 12:36:48.466180024 +0000 UTC m=+1169.408473054 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift") pod "swift-storage-0" (UID: "651edea8-3300-463c-a0fa-abd75f6056e0") : configmap "swift-ring-files" not found Dec 15 12:36:47 crc kubenswrapper[4719]: I1215 12:36:47.763520 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.504675 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:48 crc kubenswrapper[4719]: E1215 12:36:48.504949 4719 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 15 12:36:48 crc kubenswrapper[4719]: E1215 12:36:48.505160 4719 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 15 12:36:48 crc kubenswrapper[4719]: E1215 12:36:48.505228 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift podName:651edea8-3300-463c-a0fa-abd75f6056e0 nodeName:}" failed. No retries permitted until 2025-12-15 12:36:52.505208052 +0000 UTC m=+1173.447501082 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift") pod "swift-storage-0" (UID: "651edea8-3300-463c-a0fa-abd75f6056e0") : configmap "swift-ring-files" not found Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.578438 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-dcbfk"] Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.579364 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.582543 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.582733 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.583635 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.600603 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-dcbfk"] Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.708443 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rrxg\" (UniqueName: \"kubernetes.io/projected/fb01ec8f-cc85-4146-9677-344c14c2a545-kube-api-access-6rrxg\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.708483 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fb01ec8f-cc85-4146-9677-344c14c2a545-etc-swift\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.708505 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-swiftconf\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.708597 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-combined-ca-bundle\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.708637 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-scripts\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.708659 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-dispersionconf\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.708685 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-ring-data-devices\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.810441 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-combined-ca-bundle\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.810502 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-scripts\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.810526 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-dispersionconf\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.810550 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-ring-data-devices\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.810592 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rrxg\" (UniqueName: \"kubernetes.io/projected/fb01ec8f-cc85-4146-9677-344c14c2a545-kube-api-access-6rrxg\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.810610 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fb01ec8f-cc85-4146-9677-344c14c2a545-etc-swift\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.810626 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-swiftconf\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.812099 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fb01ec8f-cc85-4146-9677-344c14c2a545-etc-swift\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.812743 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-ring-data-devices\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.812802 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-scripts\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.816403 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-combined-ca-bundle\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.819252 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-swiftconf\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.824481 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-dispersionconf\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.830779 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rrxg\" (UniqueName: \"kubernetes.io/projected/fb01ec8f-cc85-4146-9677-344c14c2a545-kube-api-access-6rrxg\") pod \"swift-ring-rebalance-dcbfk\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:48 crc kubenswrapper[4719]: I1215 12:36:48.934391 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:36:49 crc kubenswrapper[4719]: I1215 12:36:49.405488 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-dcbfk"] Dec 15 12:36:49 crc kubenswrapper[4719]: W1215 12:36:49.417065 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb01ec8f_cc85_4146_9677_344c14c2a545.slice/crio-f99f2c379337022272e2d7c107ab0949711f0a04f01bd6c9add6ee429cdbbfb1 WatchSource:0}: Error finding container f99f2c379337022272e2d7c107ab0949711f0a04f01bd6c9add6ee429cdbbfb1: Status 404 returned error can't find the container with id f99f2c379337022272e2d7c107ab0949711f0a04f01bd6c9add6ee429cdbbfb1 Dec 15 12:36:50 crc kubenswrapper[4719]: I1215 12:36:50.229132 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dcbfk" event={"ID":"fb01ec8f-cc85-4146-9677-344c14c2a545","Type":"ContainerStarted","Data":"f99f2c379337022272e2d7c107ab0949711f0a04f01bd6c9add6ee429cdbbfb1"} Dec 15 12:36:52 crc kubenswrapper[4719]: I1215 12:36:52.579286 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:36:52 crc kubenswrapper[4719]: E1215 12:36:52.579783 4719 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 15 12:36:52 crc kubenswrapper[4719]: E1215 12:36:52.579797 4719 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 15 12:36:52 crc kubenswrapper[4719]: E1215 12:36:52.579890 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift podName:651edea8-3300-463c-a0fa-abd75f6056e0 nodeName:}" failed. No retries permitted until 2025-12-15 12:37:00.579839141 +0000 UTC m=+1181.522132171 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift") pod "swift-storage-0" (UID: "651edea8-3300-463c-a0fa-abd75f6056e0") : configmap "swift-ring-files" not found Dec 15 12:36:53 crc kubenswrapper[4719]: I1215 12:36:53.798018 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:36:53 crc kubenswrapper[4719]: I1215 12:36:53.857275 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-lnm79"] Dec 15 12:36:53 crc kubenswrapper[4719]: I1215 12:36:53.860456 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" podUID="aa6437c9-add6-4169-b032-406f5ddfd4e6" containerName="dnsmasq-dns" containerID="cri-o://1ab918d28fd7377927cac0863fae8fee302336c33169df244f368a1bd099f2fe" gracePeriod=10 Dec 15 12:36:53 crc kubenswrapper[4719]: I1215 12:36:53.993456 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4ht64" podUID="4c7986d1-97c7-4b8c-8b2f-594fafc24991" containerName="ovn-controller" probeResult="failure" output=< Dec 15 12:36:53 crc kubenswrapper[4719]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 15 12:36:53 crc kubenswrapper[4719]: > Dec 15 12:36:54 crc kubenswrapper[4719]: I1215 12:36:54.264611 4719 generic.go:334] "Generic (PLEG): container finished" podID="62a17473-57a0-4c96-904d-a2f68bfbc92f" containerID="0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7" exitCode=0 Dec 15 12:36:54 crc kubenswrapper[4719]: I1215 12:36:54.264735 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"62a17473-57a0-4c96-904d-a2f68bfbc92f","Type":"ContainerDied","Data":"0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7"} Dec 15 12:36:54 crc kubenswrapper[4719]: I1215 12:36:54.269647 4719 generic.go:334] "Generic (PLEG): container finished" podID="72f88cad-2122-46b3-9493-9f5f89890956" containerID="6e34719f6e20b66097cb87d45428b9ef0d5203fb97ac81e566f14c57bc6c4811" exitCode=0 Dec 15 12:36:54 crc kubenswrapper[4719]: I1215 12:36:54.269727 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"72f88cad-2122-46b3-9493-9f5f89890956","Type":"ContainerDied","Data":"6e34719f6e20b66097cb87d45428b9ef0d5203fb97ac81e566f14c57bc6c4811"} Dec 15 12:36:54 crc kubenswrapper[4719]: I1215 12:36:54.273724 4719 generic.go:334] "Generic (PLEG): container finished" podID="aa6437c9-add6-4169-b032-406f5ddfd4e6" containerID="1ab918d28fd7377927cac0863fae8fee302336c33169df244f368a1bd099f2fe" exitCode=0 Dec 15 12:36:54 crc kubenswrapper[4719]: I1215 12:36:54.273802 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" event={"ID":"aa6437c9-add6-4169-b032-406f5ddfd4e6","Type":"ContainerDied","Data":"1ab918d28fd7377927cac0863fae8fee302336c33169df244f368a1bd099f2fe"} Dec 15 12:36:56 crc kubenswrapper[4719]: I1215 12:36:56.830061 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" podUID="aa6437c9-add6-4169-b032-406f5ddfd4e6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Dec 15 12:36:58 crc kubenswrapper[4719]: I1215 12:36:58.961922 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4ht64" podUID="4c7986d1-97c7-4b8c-8b2f-594fafc24991" containerName="ovn-controller" probeResult="failure" output=< Dec 15 12:36:58 crc kubenswrapper[4719]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 15 12:36:58 crc kubenswrapper[4719]: > Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.075917 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.090935 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-hrrfr" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.344751 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4ht64-config-pp594"] Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.346226 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.352727 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.377395 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ht64-config-pp594"] Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.467012 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.529040 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-additional-scripts\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.529264 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.529337 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run-ovn\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.529369 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-scripts\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.529396 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thtx6\" (UniqueName: \"kubernetes.io/projected/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-kube-api-access-thtx6\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.529426 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-log-ovn\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.630735 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-sb\") pod \"aa6437c9-add6-4169-b032-406f5ddfd4e6\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.630841 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-config\") pod \"aa6437c9-add6-4169-b032-406f5ddfd4e6\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.631188 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc6mf\" (UniqueName: \"kubernetes.io/projected/aa6437c9-add6-4169-b032-406f5ddfd4e6-kube-api-access-fc6mf\") pod \"aa6437c9-add6-4169-b032-406f5ddfd4e6\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.631224 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-nb\") pod \"aa6437c9-add6-4169-b032-406f5ddfd4e6\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.631273 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-dns-svc\") pod \"aa6437c9-add6-4169-b032-406f5ddfd4e6\" (UID: \"aa6437c9-add6-4169-b032-406f5ddfd4e6\") " Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.631521 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-additional-scripts\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.632467 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-additional-scripts\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.632546 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.632741 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.632908 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run-ovn\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.632930 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-scripts\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.632967 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thtx6\" (UniqueName: \"kubernetes.io/projected/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-kube-api-access-thtx6\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.632980 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run-ovn\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.633024 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-log-ovn\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.632987 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-log-ovn\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.643092 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-scripts\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.654849 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa6437c9-add6-4169-b032-406f5ddfd4e6-kube-api-access-fc6mf" (OuterVolumeSpecName: "kube-api-access-fc6mf") pod "aa6437c9-add6-4169-b032-406f5ddfd4e6" (UID: "aa6437c9-add6-4169-b032-406f5ddfd4e6"). InnerVolumeSpecName "kube-api-access-fc6mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.658575 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thtx6\" (UniqueName: \"kubernetes.io/projected/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-kube-api-access-thtx6\") pod \"ovn-controller-4ht64-config-pp594\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.687638 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa6437c9-add6-4169-b032-406f5ddfd4e6" (UID: "aa6437c9-add6-4169-b032-406f5ddfd4e6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.696930 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.697799 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-config" (OuterVolumeSpecName: "config") pod "aa6437c9-add6-4169-b032-406f5ddfd4e6" (UID: "aa6437c9-add6-4169-b032-406f5ddfd4e6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.705023 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa6437c9-add6-4169-b032-406f5ddfd4e6" (UID: "aa6437c9-add6-4169-b032-406f5ddfd4e6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.725420 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa6437c9-add6-4169-b032-406f5ddfd4e6" (UID: "aa6437c9-add6-4169-b032-406f5ddfd4e6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.734354 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.734597 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc6mf\" (UniqueName: \"kubernetes.io/projected/aa6437c9-add6-4169-b032-406f5ddfd4e6-kube-api-access-fc6mf\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.734694 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.734824 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:36:59 crc kubenswrapper[4719]: I1215 12:36:59.734954 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa6437c9-add6-4169-b032-406f5ddfd4e6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.224157 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ht64-config-pp594"] Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.347302 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"72f88cad-2122-46b3-9493-9f5f89890956","Type":"ContainerStarted","Data":"5ffabd2ad1c6f430ef44131d3f76acd7216ae8256435b2ac1941539f45886823"} Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.347521 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.349725 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ht64-config-pp594" event={"ID":"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6","Type":"ContainerStarted","Data":"7f5fab9a73ec6937e9c09ab56bae8cd59d20559babc6136dfd9ee198b99d70bc"} Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.358581 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" event={"ID":"aa6437c9-add6-4169-b032-406f5ddfd4e6","Type":"ContainerDied","Data":"ac465cc236861554b8752c28506eff32f53e45c46bf932566578f422424e48a1"} Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.358624 4719 scope.go:117] "RemoveContainer" containerID="1ab918d28fd7377927cac0863fae8fee302336c33169df244f368a1bd099f2fe" Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.358721 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-lnm79" Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.369519 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nf7qt" event={"ID":"479f222a-bfc2-4257-9588-941538d75171","Type":"ContainerStarted","Data":"9d0f821426f732ad4b5c673b55eff4ecbedba7d99257789e357e7377e14f941f"} Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.371201 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"62a17473-57a0-4c96-904d-a2f68bfbc92f","Type":"ContainerStarted","Data":"25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7"} Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.371430 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.389735 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=42.382783706 podStartE2EDuration="1m23.389716554s" podCreationTimestamp="2025-12-15 12:35:37 +0000 UTC" firstStartedPulling="2025-12-15 12:35:39.427689013 +0000 UTC m=+1100.369982043" lastFinishedPulling="2025-12-15 12:36:20.434621861 +0000 UTC m=+1141.376914891" observedRunningTime="2025-12-15 12:37:00.388726253 +0000 UTC m=+1181.331019283" watchObservedRunningTime="2025-12-15 12:37:00.389716554 +0000 UTC m=+1181.332009584" Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.416584 4719 scope.go:117] "RemoveContainer" containerID="d4ab65f7f3ef39846709e553b53cbbbaf4e4a5de4fa8444b49c4af6eae027c3d" Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.423098 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=43.207871934 podStartE2EDuration="1m24.423075291s" podCreationTimestamp="2025-12-15 12:35:36 +0000 UTC" firstStartedPulling="2025-12-15 12:35:39.179216846 +0000 UTC m=+1100.121509876" lastFinishedPulling="2025-12-15 12:36:20.394420203 +0000 UTC m=+1141.336713233" observedRunningTime="2025-12-15 12:37:00.411670277 +0000 UTC m=+1181.353963307" watchObservedRunningTime="2025-12-15 12:37:00.423075291 +0000 UTC m=+1181.365368331" Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.468849 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-lnm79"] Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.494267 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-lnm79"] Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.504178 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-nf7qt" podStartSLOduration=2.492341607 podStartE2EDuration="18.504161523s" podCreationTimestamp="2025-12-15 12:36:42 +0000 UTC" firstStartedPulling="2025-12-15 12:36:43.15921678 +0000 UTC m=+1164.101509800" lastFinishedPulling="2025-12-15 12:36:59.171036696 +0000 UTC m=+1180.113329716" observedRunningTime="2025-12-15 12:37:00.44780543 +0000 UTC m=+1181.390098460" watchObservedRunningTime="2025-12-15 12:37:00.504161523 +0000 UTC m=+1181.446454553" Dec 15 12:37:00 crc kubenswrapper[4719]: I1215 12:37:00.653599 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:37:00 crc kubenswrapper[4719]: E1215 12:37:00.654074 4719 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 15 12:37:00 crc kubenswrapper[4719]: E1215 12:37:00.654099 4719 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 15 12:37:00 crc kubenswrapper[4719]: E1215 12:37:00.654172 4719 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift podName:651edea8-3300-463c-a0fa-abd75f6056e0 nodeName:}" failed. No retries permitted until 2025-12-15 12:37:16.654132276 +0000 UTC m=+1197.596425306 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift") pod "swift-storage-0" (UID: "651edea8-3300-463c-a0fa-abd75f6056e0") : configmap "swift-ring-files" not found Dec 15 12:37:01 crc kubenswrapper[4719]: I1215 12:37:01.497665 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa6437c9-add6-4169-b032-406f5ddfd4e6" path="/var/lib/kubelet/pods/aa6437c9-add6-4169-b032-406f5ddfd4e6/volumes" Dec 15 12:37:02 crc kubenswrapper[4719]: I1215 12:37:02.397044 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ht64-config-pp594" event={"ID":"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6","Type":"ContainerStarted","Data":"e76c3a5c2744508c3ceaf62081538d3fff95fc0321650867c6db547d5b5e7f7d"} Dec 15 12:37:02 crc kubenswrapper[4719]: I1215 12:37:02.425460 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4ht64-config-pp594" podStartSLOduration=3.425438609 podStartE2EDuration="3.425438609s" podCreationTimestamp="2025-12-15 12:36:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:02.415313003 +0000 UTC m=+1183.357606053" watchObservedRunningTime="2025-12-15 12:37:02.425438609 +0000 UTC m=+1183.367731649" Dec 15 12:37:03 crc kubenswrapper[4719]: I1215 12:37:03.406375 4719 generic.go:334] "Generic (PLEG): container finished" podID="92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" containerID="e76c3a5c2744508c3ceaf62081538d3fff95fc0321650867c6db547d5b5e7f7d" exitCode=0 Dec 15 12:37:03 crc kubenswrapper[4719]: I1215 12:37:03.406430 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ht64-config-pp594" event={"ID":"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6","Type":"ContainerDied","Data":"e76c3a5c2744508c3ceaf62081538d3fff95fc0321650867c6db547d5b5e7f7d"} Dec 15 12:37:03 crc kubenswrapper[4719]: I1215 12:37:03.966076 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-4ht64" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.415456 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dcbfk" event={"ID":"fb01ec8f-cc85-4146-9677-344c14c2a545","Type":"ContainerStarted","Data":"24301df43b4f42d0c8dc225274ce7832be85ca2791c515f2d16d658e14feebf7"} Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.440706 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-dcbfk" podStartSLOduration=1.824473378 podStartE2EDuration="16.440687437s" podCreationTimestamp="2025-12-15 12:36:48 +0000 UTC" firstStartedPulling="2025-12-15 12:36:49.437613957 +0000 UTC m=+1170.379906987" lastFinishedPulling="2025-12-15 12:37:04.053828016 +0000 UTC m=+1184.996121046" observedRunningTime="2025-12-15 12:37:04.437696384 +0000 UTC m=+1185.379989414" watchObservedRunningTime="2025-12-15 12:37:04.440687437 +0000 UTC m=+1185.382980467" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.718724 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.827939 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run\") pod \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828027 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-log-ovn\") pod \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828058 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run-ovn\") pod \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828124 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run" (OuterVolumeSpecName: "var-run") pod "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" (UID: "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828153 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thtx6\" (UniqueName: \"kubernetes.io/projected/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-kube-api-access-thtx6\") pod \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828190 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-additional-scripts\") pod \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828296 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-scripts\") pod \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\" (UID: \"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6\") " Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828156 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" (UID: "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828168 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" (UID: "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828700 4719 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828722 4719 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828734 4719 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.828991 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" (UID: "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.829351 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-scripts" (OuterVolumeSpecName: "scripts") pod "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" (UID: "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.832564 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-kube-api-access-thtx6" (OuterVolumeSpecName: "kube-api-access-thtx6") pod "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" (UID: "92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6"). InnerVolumeSpecName "kube-api-access-thtx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.930130 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thtx6\" (UniqueName: \"kubernetes.io/projected/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-kube-api-access-thtx6\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.930165 4719 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:04 crc kubenswrapper[4719]: I1215 12:37:04.930174 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.423467 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ht64-config-pp594" event={"ID":"92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6","Type":"ContainerDied","Data":"7f5fab9a73ec6937e9c09ab56bae8cd59d20559babc6136dfd9ee198b99d70bc"} Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.423817 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f5fab9a73ec6937e9c09ab56bae8cd59d20559babc6136dfd9ee198b99d70bc" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.423908 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ht64-config-pp594" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.548378 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4ht64-config-pp594"] Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.563257 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4ht64-config-pp594"] Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.691734 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4ht64-config-j447z"] Dec 15 12:37:05 crc kubenswrapper[4719]: E1215 12:37:05.692470 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6437c9-add6-4169-b032-406f5ddfd4e6" containerName="init" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.692564 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6437c9-add6-4169-b032-406f5ddfd4e6" containerName="init" Dec 15 12:37:05 crc kubenswrapper[4719]: E1215 12:37:05.692657 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" containerName="ovn-config" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.692731 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" containerName="ovn-config" Dec 15 12:37:05 crc kubenswrapper[4719]: E1215 12:37:05.692829 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6437c9-add6-4169-b032-406f5ddfd4e6" containerName="dnsmasq-dns" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.692924 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6437c9-add6-4169-b032-406f5ddfd4e6" containerName="dnsmasq-dns" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.693169 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa6437c9-add6-4169-b032-406f5ddfd4e6" containerName="dnsmasq-dns" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.693260 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" containerName="ovn-config" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.693943 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.702480 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.704206 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ht64-config-j447z"] Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.844753 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-scripts\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.845067 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run-ovn\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.845207 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-additional-scripts\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.845352 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vk4s\" (UniqueName: \"kubernetes.io/projected/b192e21c-1384-4766-a0a4-3fe22437e6a0-kube-api-access-4vk4s\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.845467 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.845592 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-log-ovn\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.947018 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run-ovn\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.947319 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run-ovn\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.947342 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-scripts\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.947555 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-additional-scripts\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.947735 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vk4s\" (UniqueName: \"kubernetes.io/projected/b192e21c-1384-4766-a0a4-3fe22437e6a0-kube-api-access-4vk4s\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.948225 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-additional-scripts\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.948352 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.948377 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.948576 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-log-ovn\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.948618 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-log-ovn\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.949422 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-scripts\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:05 crc kubenswrapper[4719]: I1215 12:37:05.968710 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vk4s\" (UniqueName: \"kubernetes.io/projected/b192e21c-1384-4766-a0a4-3fe22437e6a0-kube-api-access-4vk4s\") pod \"ovn-controller-4ht64-config-j447z\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:06 crc kubenswrapper[4719]: I1215 12:37:06.010484 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:06 crc kubenswrapper[4719]: I1215 12:37:06.532537 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ht64-config-j447z"] Dec 15 12:37:07 crc kubenswrapper[4719]: I1215 12:37:07.447134 4719 generic.go:334] "Generic (PLEG): container finished" podID="b192e21c-1384-4766-a0a4-3fe22437e6a0" containerID="afeddce05a6681af855875ccab7c2a377353193f66c8804e0f1ec4be1dd21b07" exitCode=0 Dec 15 12:37:07 crc kubenswrapper[4719]: I1215 12:37:07.447277 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ht64-config-j447z" event={"ID":"b192e21c-1384-4766-a0a4-3fe22437e6a0","Type":"ContainerDied","Data":"afeddce05a6681af855875ccab7c2a377353193f66c8804e0f1ec4be1dd21b07"} Dec 15 12:37:07 crc kubenswrapper[4719]: I1215 12:37:07.447472 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ht64-config-j447z" event={"ID":"b192e21c-1384-4766-a0a4-3fe22437e6a0","Type":"ContainerStarted","Data":"7b8a6b387bf3e0a86e2b10c987d3493817c752144bf9acdd5ab3f7958136088e"} Dec 15 12:37:07 crc kubenswrapper[4719]: I1215 12:37:07.488013 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6" path="/var/lib/kubelet/pods/92e2c91d-20cf-4570-b5b4-56bdbb2cc5e6/volumes" Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.812877 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.916248 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run\") pod \"b192e21c-1384-4766-a0a4-3fe22437e6a0\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.916411 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vk4s\" (UniqueName: \"kubernetes.io/projected/b192e21c-1384-4766-a0a4-3fe22437e6a0-kube-api-access-4vk4s\") pod \"b192e21c-1384-4766-a0a4-3fe22437e6a0\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.916399 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run" (OuterVolumeSpecName: "var-run") pod "b192e21c-1384-4766-a0a4-3fe22437e6a0" (UID: "b192e21c-1384-4766-a0a4-3fe22437e6a0"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.916442 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-log-ovn\") pod \"b192e21c-1384-4766-a0a4-3fe22437e6a0\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.916486 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run-ovn\") pod \"b192e21c-1384-4766-a0a4-3fe22437e6a0\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.916518 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-scripts\") pod \"b192e21c-1384-4766-a0a4-3fe22437e6a0\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.916544 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-additional-scripts\") pod \"b192e21c-1384-4766-a0a4-3fe22437e6a0\" (UID: \"b192e21c-1384-4766-a0a4-3fe22437e6a0\") " Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.916755 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b192e21c-1384-4766-a0a4-3fe22437e6a0" (UID: "b192e21c-1384-4766-a0a4-3fe22437e6a0"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.916786 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b192e21c-1384-4766-a0a4-3fe22437e6a0" (UID: "b192e21c-1384-4766-a0a4-3fe22437e6a0"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.917173 4719 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.917195 4719 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.917204 4719 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b192e21c-1384-4766-a0a4-3fe22437e6a0-var-run\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.917591 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b192e21c-1384-4766-a0a4-3fe22437e6a0" (UID: "b192e21c-1384-4766-a0a4-3fe22437e6a0"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.917632 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-scripts" (OuterVolumeSpecName: "scripts") pod "b192e21c-1384-4766-a0a4-3fe22437e6a0" (UID: "b192e21c-1384-4766-a0a4-3fe22437e6a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:08 crc kubenswrapper[4719]: I1215 12:37:08.939109 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b192e21c-1384-4766-a0a4-3fe22437e6a0-kube-api-access-4vk4s" (OuterVolumeSpecName: "kube-api-access-4vk4s") pod "b192e21c-1384-4766-a0a4-3fe22437e6a0" (UID: "b192e21c-1384-4766-a0a4-3fe22437e6a0"). InnerVolumeSpecName "kube-api-access-4vk4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:09 crc kubenswrapper[4719]: I1215 12:37:09.018700 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:09 crc kubenswrapper[4719]: I1215 12:37:09.018732 4719 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b192e21c-1384-4766-a0a4-3fe22437e6a0-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:09 crc kubenswrapper[4719]: I1215 12:37:09.018742 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vk4s\" (UniqueName: \"kubernetes.io/projected/b192e21c-1384-4766-a0a4-3fe22437e6a0-kube-api-access-4vk4s\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:09 crc kubenswrapper[4719]: I1215 12:37:09.466183 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ht64-config-j447z" event={"ID":"b192e21c-1384-4766-a0a4-3fe22437e6a0","Type":"ContainerDied","Data":"7b8a6b387bf3e0a86e2b10c987d3493817c752144bf9acdd5ab3f7958136088e"} Dec 15 12:37:09 crc kubenswrapper[4719]: I1215 12:37:09.466534 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b8a6b387bf3e0a86e2b10c987d3493817c752144bf9acdd5ab3f7958136088e" Dec 15 12:37:09 crc kubenswrapper[4719]: I1215 12:37:09.466212 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ht64-config-j447z" Dec 15 12:37:09 crc kubenswrapper[4719]: I1215 12:37:09.898454 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4ht64-config-j447z"] Dec 15 12:37:09 crc kubenswrapper[4719]: I1215 12:37:09.906295 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4ht64-config-j447z"] Dec 15 12:37:10 crc kubenswrapper[4719]: I1215 12:37:10.473932 4719 generic.go:334] "Generic (PLEG): container finished" podID="479f222a-bfc2-4257-9588-941538d75171" containerID="9d0f821426f732ad4b5c673b55eff4ecbedba7d99257789e357e7377e14f941f" exitCode=0 Dec 15 12:37:10 crc kubenswrapper[4719]: I1215 12:37:10.473993 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nf7qt" event={"ID":"479f222a-bfc2-4257-9588-941538d75171","Type":"ContainerDied","Data":"9d0f821426f732ad4b5c673b55eff4ecbedba7d99257789e357e7377e14f941f"} Dec 15 12:37:11 crc kubenswrapper[4719]: I1215 12:37:11.487454 4719 generic.go:334] "Generic (PLEG): container finished" podID="fb01ec8f-cc85-4146-9677-344c14c2a545" containerID="24301df43b4f42d0c8dc225274ce7832be85ca2791c515f2d16d658e14feebf7" exitCode=0 Dec 15 12:37:11 crc kubenswrapper[4719]: I1215 12:37:11.492498 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b192e21c-1384-4766-a0a4-3fe22437e6a0" path="/var/lib/kubelet/pods/b192e21c-1384-4766-a0a4-3fe22437e6a0/volumes" Dec 15 12:37:11 crc kubenswrapper[4719]: I1215 12:37:11.493419 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dcbfk" event={"ID":"fb01ec8f-cc85-4146-9677-344c14c2a545","Type":"ContainerDied","Data":"24301df43b4f42d0c8dc225274ce7832be85ca2791c515f2d16d658e14feebf7"} Dec 15 12:37:11 crc kubenswrapper[4719]: I1215 12:37:11.891379 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nf7qt" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.061556 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-combined-ca-bundle\") pod \"479f222a-bfc2-4257-9588-941538d75171\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.061608 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-db-sync-config-data\") pod \"479f222a-bfc2-4257-9588-941538d75171\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.061668 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgqq8\" (UniqueName: \"kubernetes.io/projected/479f222a-bfc2-4257-9588-941538d75171-kube-api-access-jgqq8\") pod \"479f222a-bfc2-4257-9588-941538d75171\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.061836 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-config-data\") pod \"479f222a-bfc2-4257-9588-941538d75171\" (UID: \"479f222a-bfc2-4257-9588-941538d75171\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.066640 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/479f222a-bfc2-4257-9588-941538d75171-kube-api-access-jgqq8" (OuterVolumeSpecName: "kube-api-access-jgqq8") pod "479f222a-bfc2-4257-9588-941538d75171" (UID: "479f222a-bfc2-4257-9588-941538d75171"). InnerVolumeSpecName "kube-api-access-jgqq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.070579 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "479f222a-bfc2-4257-9588-941538d75171" (UID: "479f222a-bfc2-4257-9588-941538d75171"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.085116 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "479f222a-bfc2-4257-9588-941538d75171" (UID: "479f222a-bfc2-4257-9588-941538d75171"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.104659 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-config-data" (OuterVolumeSpecName: "config-data") pod "479f222a-bfc2-4257-9588-941538d75171" (UID: "479f222a-bfc2-4257-9588-941538d75171"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.163836 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.163914 4719 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.163934 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgqq8\" (UniqueName: \"kubernetes.io/projected/479f222a-bfc2-4257-9588-941538d75171-kube-api-access-jgqq8\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.163951 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/479f222a-bfc2-4257-9588-941538d75171-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.498148 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nf7qt" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.498222 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nf7qt" event={"ID":"479f222a-bfc2-4257-9588-941538d75171","Type":"ContainerDied","Data":"5479d7b82227bac705b45f45fae85af34a996ecafcdcd3f96acab69b64bb1b58"} Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.499255 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5479d7b82227bac705b45f45fae85af34a996ecafcdcd3f96acab69b64bb1b58" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.783274 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.875976 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-scripts\") pod \"fb01ec8f-cc85-4146-9677-344c14c2a545\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.876046 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-swiftconf\") pod \"fb01ec8f-cc85-4146-9677-344c14c2a545\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.876079 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-dispersionconf\") pod \"fb01ec8f-cc85-4146-9677-344c14c2a545\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.876108 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fb01ec8f-cc85-4146-9677-344c14c2a545-etc-swift\") pod \"fb01ec8f-cc85-4146-9677-344c14c2a545\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.877197 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb01ec8f-cc85-4146-9677-344c14c2a545-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "fb01ec8f-cc85-4146-9677-344c14c2a545" (UID: "fb01ec8f-cc85-4146-9677-344c14c2a545"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.876158 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-combined-ca-bundle\") pod \"fb01ec8f-cc85-4146-9677-344c14c2a545\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.877292 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-ring-data-devices\") pod \"fb01ec8f-cc85-4146-9677-344c14c2a545\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.878399 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "fb01ec8f-cc85-4146-9677-344c14c2a545" (UID: "fb01ec8f-cc85-4146-9677-344c14c2a545"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.878443 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rrxg\" (UniqueName: \"kubernetes.io/projected/fb01ec8f-cc85-4146-9677-344c14c2a545-kube-api-access-6rrxg\") pod \"fb01ec8f-cc85-4146-9677-344c14c2a545\" (UID: \"fb01ec8f-cc85-4146-9677-344c14c2a545\") " Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.878739 4719 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fb01ec8f-cc85-4146-9677-344c14c2a545-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.878753 4719 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.917183 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-scripts" (OuterVolumeSpecName: "scripts") pod "fb01ec8f-cc85-4146-9677-344c14c2a545" (UID: "fb01ec8f-cc85-4146-9677-344c14c2a545"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.921037 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "fb01ec8f-cc85-4146-9677-344c14c2a545" (UID: "fb01ec8f-cc85-4146-9677-344c14c2a545"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.926134 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb01ec8f-cc85-4146-9677-344c14c2a545-kube-api-access-6rrxg" (OuterVolumeSpecName: "kube-api-access-6rrxg") pod "fb01ec8f-cc85-4146-9677-344c14c2a545" (UID: "fb01ec8f-cc85-4146-9677-344c14c2a545"). InnerVolumeSpecName "kube-api-access-6rrxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.962702 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb01ec8f-cc85-4146-9677-344c14c2a545" (UID: "fb01ec8f-cc85-4146-9677-344c14c2a545"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.981415 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb01ec8f-cc85-4146-9677-344c14c2a545-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.981447 4719 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.981462 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:12 crc kubenswrapper[4719]: I1215 12:37:12.981472 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rrxg\" (UniqueName: \"kubernetes.io/projected/fb01ec8f-cc85-4146-9677-344c14c2a545-kube-api-access-6rrxg\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.001126 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "fb01ec8f-cc85-4146-9677-344c14c2a545" (UID: "fb01ec8f-cc85-4146-9677-344c14c2a545"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.011461 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-m9j55"] Dec 15 12:37:13 crc kubenswrapper[4719]: E1215 12:37:13.012008 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb01ec8f-cc85-4146-9677-344c14c2a545" containerName="swift-ring-rebalance" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.012031 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb01ec8f-cc85-4146-9677-344c14c2a545" containerName="swift-ring-rebalance" Dec 15 12:37:13 crc kubenswrapper[4719]: E1215 12:37:13.012087 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b192e21c-1384-4766-a0a4-3fe22437e6a0" containerName="ovn-config" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.012097 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b192e21c-1384-4766-a0a4-3fe22437e6a0" containerName="ovn-config" Dec 15 12:37:13 crc kubenswrapper[4719]: E1215 12:37:13.012111 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="479f222a-bfc2-4257-9588-941538d75171" containerName="glance-db-sync" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.012119 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="479f222a-bfc2-4257-9588-941538d75171" containerName="glance-db-sync" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.012341 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="479f222a-bfc2-4257-9588-941538d75171" containerName="glance-db-sync" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.012367 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="b192e21c-1384-4766-a0a4-3fe22437e6a0" containerName="ovn-config" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.012378 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb01ec8f-cc85-4146-9677-344c14c2a545" containerName="swift-ring-rebalance" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.013510 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.026401 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-m9j55"] Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.084360 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-config\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.084996 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.085136 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.085239 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx5bh\" (UniqueName: \"kubernetes.io/projected/9bb0baa8-ce19-4923-9e24-66933db4c618-kube-api-access-jx5bh\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.085381 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.085528 4719 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fb01ec8f-cc85-4146-9677-344c14c2a545-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.186417 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-config\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.186558 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.186596 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.186618 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx5bh\" (UniqueName: \"kubernetes.io/projected/9bb0baa8-ce19-4923-9e24-66933db4c618-kube-api-access-jx5bh\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.186640 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.187394 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-config\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.187426 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.187518 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.188351 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.221892 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx5bh\" (UniqueName: \"kubernetes.io/projected/9bb0baa8-ce19-4923-9e24-66933db4c618-kube-api-access-jx5bh\") pod \"dnsmasq-dns-5b946c75cc-m9j55\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.384482 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.529933 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dcbfk" event={"ID":"fb01ec8f-cc85-4146-9677-344c14c2a545","Type":"ContainerDied","Data":"f99f2c379337022272e2d7c107ab0949711f0a04f01bd6c9add6ee429cdbbfb1"} Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.529977 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f99f2c379337022272e2d7c107ab0949711f0a04f01bd6c9add6ee429cdbbfb1" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.530039 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dcbfk" Dec 15 12:37:13 crc kubenswrapper[4719]: I1215 12:37:13.885394 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-m9j55"] Dec 15 12:37:14 crc kubenswrapper[4719]: I1215 12:37:14.537649 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" event={"ID":"9bb0baa8-ce19-4923-9e24-66933db4c618","Type":"ContainerStarted","Data":"2b8b726944c1c2880bff5013ca1db9803be59ad45707d31ccffb449d2bcf64be"} Dec 15 12:37:15 crc kubenswrapper[4719]: I1215 12:37:15.553501 4719 generic.go:334] "Generic (PLEG): container finished" podID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerID="29cbf0603161715ba41954c2791f0169f0df10391331aac7a7dae583f4bd69af" exitCode=0 Dec 15 12:37:15 crc kubenswrapper[4719]: I1215 12:37:15.553645 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" event={"ID":"9bb0baa8-ce19-4923-9e24-66933db4c618","Type":"ContainerDied","Data":"29cbf0603161715ba41954c2791f0169f0df10391331aac7a7dae583f4bd69af"} Dec 15 12:37:16 crc kubenswrapper[4719]: I1215 12:37:16.579089 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" event={"ID":"9bb0baa8-ce19-4923-9e24-66933db4c618","Type":"ContainerStarted","Data":"0e96ac6e3ea981f737bec1c9446e922db723f7ab9152cce21ff8a3f7e3d80063"} Dec 15 12:37:16 crc kubenswrapper[4719]: I1215 12:37:16.579940 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:16 crc kubenswrapper[4719]: I1215 12:37:16.601371 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" podStartSLOduration=4.601350266 podStartE2EDuration="4.601350266s" podCreationTimestamp="2025-12-15 12:37:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:16.595832485 +0000 UTC m=+1197.538125515" watchObservedRunningTime="2025-12-15 12:37:16.601350266 +0000 UTC m=+1197.543643296" Dec 15 12:37:16 crc kubenswrapper[4719]: I1215 12:37:16.744707 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:37:16 crc kubenswrapper[4719]: I1215 12:37:16.750445 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/651edea8-3300-463c-a0fa-abd75f6056e0-etc-swift\") pod \"swift-storage-0\" (UID: \"651edea8-3300-463c-a0fa-abd75f6056e0\") " pod="openstack/swift-storage-0" Dec 15 12:37:16 crc kubenswrapper[4719]: I1215 12:37:16.904767 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 15 12:37:17 crc kubenswrapper[4719]: I1215 12:37:17.501350 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 15 12:37:17 crc kubenswrapper[4719]: W1215 12:37:17.511968 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod651edea8_3300_463c_a0fa_abd75f6056e0.slice/crio-875e25706b0327bb490c65c72ad1e4837dcb6a2a4d5ae940fa8dd78e73c5cd55 WatchSource:0}: Error finding container 875e25706b0327bb490c65c72ad1e4837dcb6a2a4d5ae940fa8dd78e73c5cd55: Status 404 returned error can't find the container with id 875e25706b0327bb490c65c72ad1e4837dcb6a2a4d5ae940fa8dd78e73c5cd55 Dec 15 12:37:17 crc kubenswrapper[4719]: I1215 12:37:17.587265 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"875e25706b0327bb490c65c72ad1e4837dcb6a2a4d5ae940fa8dd78e73c5cd55"} Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.203818 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.596556 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-lgf62"] Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.597744 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lgf62" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.604669 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-6f1e-account-create-update-bps6c"] Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.606146 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6f1e-account-create-update-bps6c" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.607707 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.628294 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6f1e-account-create-update-bps6c"] Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.637425 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lgf62"] Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.686597 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz54b\" (UniqueName: \"kubernetes.io/projected/c0e6d128-ab9b-4a15-bef7-c986942dc16f-kube-api-access-vz54b\") pod \"cinder-6f1e-account-create-update-bps6c\" (UID: \"c0e6d128-ab9b-4a15-bef7-c986942dc16f\") " pod="openstack/cinder-6f1e-account-create-update-bps6c" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.686646 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8kpw\" (UniqueName: \"kubernetes.io/projected/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-kube-api-access-r8kpw\") pod \"cinder-db-create-lgf62\" (UID: \"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1\") " pod="openstack/cinder-db-create-lgf62" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.686711 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-operator-scripts\") pod \"cinder-db-create-lgf62\" (UID: \"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1\") " pod="openstack/cinder-db-create-lgf62" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.686776 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0e6d128-ab9b-4a15-bef7-c986942dc16f-operator-scripts\") pod \"cinder-6f1e-account-create-update-bps6c\" (UID: \"c0e6d128-ab9b-4a15-bef7-c986942dc16f\") " pod="openstack/cinder-6f1e-account-create-update-bps6c" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.696488 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-75qg9"] Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.697647 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-75qg9" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.716601 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-75qg9"] Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.729784 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-539a-account-create-update-zqll7"] Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.730783 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-539a-account-create-update-zqll7" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.735119 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.738870 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-539a-account-create-update-zqll7"] Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.797845 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0e6d128-ab9b-4a15-bef7-c986942dc16f-operator-scripts\") pod \"cinder-6f1e-account-create-update-bps6c\" (UID: \"c0e6d128-ab9b-4a15-bef7-c986942dc16f\") " pod="openstack/cinder-6f1e-account-create-update-bps6c" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.798004 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz54b\" (UniqueName: \"kubernetes.io/projected/c0e6d128-ab9b-4a15-bef7-c986942dc16f-kube-api-access-vz54b\") pod \"cinder-6f1e-account-create-update-bps6c\" (UID: \"c0e6d128-ab9b-4a15-bef7-c986942dc16f\") " pod="openstack/cinder-6f1e-account-create-update-bps6c" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.798061 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8kpw\" (UniqueName: \"kubernetes.io/projected/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-kube-api-access-r8kpw\") pod \"cinder-db-create-lgf62\" (UID: \"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1\") " pod="openstack/cinder-db-create-lgf62" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.798104 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-operator-scripts\") pod \"cinder-db-create-lgf62\" (UID: \"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1\") " pod="openstack/cinder-db-create-lgf62" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.805816 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-operator-scripts\") pod \"cinder-db-create-lgf62\" (UID: \"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1\") " pod="openstack/cinder-db-create-lgf62" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.808638 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0e6d128-ab9b-4a15-bef7-c986942dc16f-operator-scripts\") pod \"cinder-6f1e-account-create-update-bps6c\" (UID: \"c0e6d128-ab9b-4a15-bef7-c986942dc16f\") " pod="openstack/cinder-6f1e-account-create-update-bps6c" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.846296 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz54b\" (UniqueName: \"kubernetes.io/projected/c0e6d128-ab9b-4a15-bef7-c986942dc16f-kube-api-access-vz54b\") pod \"cinder-6f1e-account-create-update-bps6c\" (UID: \"c0e6d128-ab9b-4a15-bef7-c986942dc16f\") " pod="openstack/cinder-6f1e-account-create-update-bps6c" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.865967 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8kpw\" (UniqueName: \"kubernetes.io/projected/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-kube-api-access-r8kpw\") pod \"cinder-db-create-lgf62\" (UID: \"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1\") " pod="openstack/cinder-db-create-lgf62" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.900908 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0381936c-f528-4662-8a86-558c341065c0-operator-scripts\") pod \"barbican-db-create-75qg9\" (UID: \"0381936c-f528-4662-8a86-558c341065c0\") " pod="openstack/barbican-db-create-75qg9" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.901151 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp6wt\" (UniqueName: \"kubernetes.io/projected/0381936c-f528-4662-8a86-558c341065c0-kube-api-access-hp6wt\") pod \"barbican-db-create-75qg9\" (UID: \"0381936c-f528-4662-8a86-558c341065c0\") " pod="openstack/barbican-db-create-75qg9" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.901299 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83819057-000b-4f9e-b8fc-3478bc403f4e-operator-scripts\") pod \"barbican-539a-account-create-update-zqll7\" (UID: \"83819057-000b-4f9e-b8fc-3478bc403f4e\") " pod="openstack/barbican-539a-account-create-update-zqll7" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.901424 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj725\" (UniqueName: \"kubernetes.io/projected/83819057-000b-4f9e-b8fc-3478bc403f4e-kube-api-access-wj725\") pod \"barbican-539a-account-create-update-zqll7\" (UID: \"83819057-000b-4f9e-b8fc-3478bc403f4e\") " pod="openstack/barbican-539a-account-create-update-zqll7" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.905507 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.926505 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lgf62" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.940906 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d2d3-account-create-update-8lpdk"] Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.942622 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6f1e-account-create-update-bps6c" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.942740 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d2d3-account-create-update-8lpdk" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.944673 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.975359 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-tcmf2"] Dec 15 12:37:18 crc kubenswrapper[4719]: I1215 12:37:18.976764 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tcmf2" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:18.999989 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d2d3-account-create-update-8lpdk"] Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.002875 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0381936c-f528-4662-8a86-558c341065c0-operator-scripts\") pod \"barbican-db-create-75qg9\" (UID: \"0381936c-f528-4662-8a86-558c341065c0\") " pod="openstack/barbican-db-create-75qg9" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.002925 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp6wt\" (UniqueName: \"kubernetes.io/projected/0381936c-f528-4662-8a86-558c341065c0-kube-api-access-hp6wt\") pod \"barbican-db-create-75qg9\" (UID: \"0381936c-f528-4662-8a86-558c341065c0\") " pod="openstack/barbican-db-create-75qg9" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.003008 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83819057-000b-4f9e-b8fc-3478bc403f4e-operator-scripts\") pod \"barbican-539a-account-create-update-zqll7\" (UID: \"83819057-000b-4f9e-b8fc-3478bc403f4e\") " pod="openstack/barbican-539a-account-create-update-zqll7" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.003080 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj725\" (UniqueName: \"kubernetes.io/projected/83819057-000b-4f9e-b8fc-3478bc403f4e-kube-api-access-wj725\") pod \"barbican-539a-account-create-update-zqll7\" (UID: \"83819057-000b-4f9e-b8fc-3478bc403f4e\") " pod="openstack/barbican-539a-account-create-update-zqll7" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.009221 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0381936c-f528-4662-8a86-558c341065c0-operator-scripts\") pod \"barbican-db-create-75qg9\" (UID: \"0381936c-f528-4662-8a86-558c341065c0\") " pod="openstack/barbican-db-create-75qg9" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.010473 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83819057-000b-4f9e-b8fc-3478bc403f4e-operator-scripts\") pod \"barbican-539a-account-create-update-zqll7\" (UID: \"83819057-000b-4f9e-b8fc-3478bc403f4e\") " pod="openstack/barbican-539a-account-create-update-zqll7" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.018883 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tcmf2"] Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.047893 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-c6wxw"] Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.048957 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.051959 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.052106 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.052216 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kdx8k" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.062636 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.074542 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp6wt\" (UniqueName: \"kubernetes.io/projected/0381936c-f528-4662-8a86-558c341065c0-kube-api-access-hp6wt\") pod \"barbican-db-create-75qg9\" (UID: \"0381936c-f528-4662-8a86-558c341065c0\") " pod="openstack/barbican-db-create-75qg9" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.091436 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj725\" (UniqueName: \"kubernetes.io/projected/83819057-000b-4f9e-b8fc-3478bc403f4e-kube-api-access-wj725\") pod \"barbican-539a-account-create-update-zqll7\" (UID: \"83819057-000b-4f9e-b8fc-3478bc403f4e\") " pod="openstack/barbican-539a-account-create-update-zqll7" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.099976 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c6wxw"] Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.104789 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-operator-scripts\") pod \"neutron-d2d3-account-create-update-8lpdk\" (UID: \"fdc568d9-e75c-4f9b-a426-4dbc175a9f44\") " pod="openstack/neutron-d2d3-account-create-update-8lpdk" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.105024 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqv7s\" (UniqueName: \"kubernetes.io/projected/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-kube-api-access-cqv7s\") pod \"neutron-d2d3-account-create-update-8lpdk\" (UID: \"fdc568d9-e75c-4f9b-a426-4dbc175a9f44\") " pod="openstack/neutron-d2d3-account-create-update-8lpdk" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.105098 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-operator-scripts\") pod \"neutron-db-create-tcmf2\" (UID: \"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752\") " pod="openstack/neutron-db-create-tcmf2" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.105194 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf92m\" (UniqueName: \"kubernetes.io/projected/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-kube-api-access-zf92m\") pod \"neutron-db-create-tcmf2\" (UID: \"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752\") " pod="openstack/neutron-db-create-tcmf2" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.208608 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k994q\" (UniqueName: \"kubernetes.io/projected/4c64d8b6-1005-484e-b068-95d462167a8c-kube-api-access-k994q\") pod \"keystone-db-sync-c6wxw\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.208710 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqv7s\" (UniqueName: \"kubernetes.io/projected/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-kube-api-access-cqv7s\") pod \"neutron-d2d3-account-create-update-8lpdk\" (UID: \"fdc568d9-e75c-4f9b-a426-4dbc175a9f44\") " pod="openstack/neutron-d2d3-account-create-update-8lpdk" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.208734 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-operator-scripts\") pod \"neutron-db-create-tcmf2\" (UID: \"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752\") " pod="openstack/neutron-db-create-tcmf2" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.208757 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-config-data\") pod \"keystone-db-sync-c6wxw\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.208785 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-combined-ca-bundle\") pod \"keystone-db-sync-c6wxw\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.208839 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf92m\" (UniqueName: \"kubernetes.io/projected/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-kube-api-access-zf92m\") pod \"neutron-db-create-tcmf2\" (UID: \"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752\") " pod="openstack/neutron-db-create-tcmf2" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.209183 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-operator-scripts\") pod \"neutron-d2d3-account-create-update-8lpdk\" (UID: \"fdc568d9-e75c-4f9b-a426-4dbc175a9f44\") " pod="openstack/neutron-d2d3-account-create-update-8lpdk" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.209952 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-operator-scripts\") pod \"neutron-d2d3-account-create-update-8lpdk\" (UID: \"fdc568d9-e75c-4f9b-a426-4dbc175a9f44\") " pod="openstack/neutron-d2d3-account-create-update-8lpdk" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.210281 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-operator-scripts\") pod \"neutron-db-create-tcmf2\" (UID: \"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752\") " pod="openstack/neutron-db-create-tcmf2" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.229920 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf92m\" (UniqueName: \"kubernetes.io/projected/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-kube-api-access-zf92m\") pod \"neutron-db-create-tcmf2\" (UID: \"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752\") " pod="openstack/neutron-db-create-tcmf2" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.231533 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqv7s\" (UniqueName: \"kubernetes.io/projected/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-kube-api-access-cqv7s\") pod \"neutron-d2d3-account-create-update-8lpdk\" (UID: \"fdc568d9-e75c-4f9b-a426-4dbc175a9f44\") " pod="openstack/neutron-d2d3-account-create-update-8lpdk" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.273469 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d2d3-account-create-update-8lpdk" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.307631 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tcmf2" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.310820 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-config-data\") pod \"keystone-db-sync-c6wxw\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.310894 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-combined-ca-bundle\") pod \"keystone-db-sync-c6wxw\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.311230 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k994q\" (UniqueName: \"kubernetes.io/projected/4c64d8b6-1005-484e-b068-95d462167a8c-kube-api-access-k994q\") pod \"keystone-db-sync-c6wxw\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.314753 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-config-data\") pod \"keystone-db-sync-c6wxw\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.315432 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-combined-ca-bundle\") pod \"keystone-db-sync-c6wxw\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.324355 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-75qg9" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.331368 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k994q\" (UniqueName: \"kubernetes.io/projected/4c64d8b6-1005-484e-b068-95d462167a8c-kube-api-access-k994q\") pod \"keystone-db-sync-c6wxw\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.363220 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-539a-account-create-update-zqll7" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.473690 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kdx8k" Dec 15 12:37:19 crc kubenswrapper[4719]: I1215 12:37:19.485819 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.161431 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-539a-account-create-update-zqll7"] Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.269047 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.549018 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6f1e-account-create-update-bps6c"] Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.575842 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d2d3-account-create-update-8lpdk"] Dec 15 12:37:20 crc kubenswrapper[4719]: W1215 12:37:20.576116 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0e6d128_ab9b_4a15_bef7_c986942dc16f.slice/crio-25ec6464b837bfff94aaef1302e567575f14dc77711f9f028071fcfe941a1e61 WatchSource:0}: Error finding container 25ec6464b837bfff94aaef1302e567575f14dc77711f9f028071fcfe941a1e61: Status 404 returned error can't find the container with id 25ec6464b837bfff94aaef1302e567575f14dc77711f9f028071fcfe941a1e61 Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.582434 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 15 12:37:20 crc kubenswrapper[4719]: W1215 12:37:20.592623 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdc568d9_e75c_4f9b_a426_4dbc175a9f44.slice/crio-554b9096284f62436d80500e92bec603f32d5e1664fc475fc25aaae3ab60aca9 WatchSource:0}: Error finding container 554b9096284f62436d80500e92bec603f32d5e1664fc475fc25aaae3ab60aca9: Status 404 returned error can't find the container with id 554b9096284f62436d80500e92bec603f32d5e1664fc475fc25aaae3ab60aca9 Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.612968 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.621754 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d2d3-account-create-update-8lpdk" event={"ID":"fdc568d9-e75c-4f9b-a426-4dbc175a9f44","Type":"ContainerStarted","Data":"554b9096284f62436d80500e92bec603f32d5e1664fc475fc25aaae3ab60aca9"} Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.631006 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6f1e-account-create-update-bps6c" event={"ID":"c0e6d128-ab9b-4a15-bef7-c986942dc16f","Type":"ContainerStarted","Data":"25ec6464b837bfff94aaef1302e567575f14dc77711f9f028071fcfe941a1e61"} Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.632529 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-539a-account-create-update-zqll7" event={"ID":"83819057-000b-4f9e-b8fc-3478bc403f4e","Type":"ContainerStarted","Data":"fb301aeb8662ca386a1b946e9d89d0d5126b0c82f9f845b0c7de32fa368ebbca"} Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.632550 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-539a-account-create-update-zqll7" event={"ID":"83819057-000b-4f9e-b8fc-3478bc403f4e","Type":"ContainerStarted","Data":"0064de5b30526464acc268f4f5bcc296566bac160d7527c71a1c266bea08577c"} Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.639598 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"520a48ef167a441cc8b225e3f5d22fdeefbc85a8fb904bc209a7c64e5c20e3ed"} Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.639634 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"978cdd78f7baaa4c08c7879e7df90ea18e2847f78a3fa809fb0373eaeba5f684"} Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.651130 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-539a-account-create-update-zqll7" podStartSLOduration=2.651114042 podStartE2EDuration="2.651114042s" podCreationTimestamp="2025-12-15 12:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:20.646301872 +0000 UTC m=+1201.588594902" watchObservedRunningTime="2025-12-15 12:37:20.651114042 +0000 UTC m=+1201.593407072" Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.675410 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c6wxw"] Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.686442 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-lgf62"] Dec 15 12:37:20 crc kubenswrapper[4719]: W1215 12:37:20.709178 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2ec6917_fd83_4b5b_b904_6a0e4d453cf1.slice/crio-c9ed5081a49c8d75ec7de04fa383fefa1599366d248181b2be27eab54b5f12e3 WatchSource:0}: Error finding container c9ed5081a49c8d75ec7de04fa383fefa1599366d248181b2be27eab54b5f12e3: Status 404 returned error can't find the container with id c9ed5081a49c8d75ec7de04fa383fefa1599366d248181b2be27eab54b5f12e3 Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.791041 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-75qg9"] Dec 15 12:37:20 crc kubenswrapper[4719]: I1215 12:37:20.828785 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tcmf2"] Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.652520 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"87979053270a00242097a298815f8ef72cc5b589a8d5911f1ac8f51f3b0e8957"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.652789 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"f3ddd2069aa5606d0e155550f9fce226f37f08834b8dc48b0a54de5f8f5e9524"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.655428 4719 generic.go:334] "Generic (PLEG): container finished" podID="e2ec6917-fd83-4b5b-b904-6a0e4d453cf1" containerID="3616a57fe36a9e7aea2d305291ee91e9b90b8da3223af440ba8b48459410b4ec" exitCode=0 Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.655510 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lgf62" event={"ID":"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1","Type":"ContainerDied","Data":"3616a57fe36a9e7aea2d305291ee91e9b90b8da3223af440ba8b48459410b4ec"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.655537 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lgf62" event={"ID":"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1","Type":"ContainerStarted","Data":"c9ed5081a49c8d75ec7de04fa383fefa1599366d248181b2be27eab54b5f12e3"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.661006 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-75qg9" event={"ID":"0381936c-f528-4662-8a86-558c341065c0","Type":"ContainerStarted","Data":"5f094fe3136a3f3ce0225bc95beac5be77d50f6bddf8229c8d1be2f03cab6b51"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.661048 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-75qg9" event={"ID":"0381936c-f528-4662-8a86-558c341065c0","Type":"ContainerStarted","Data":"c91bf3e196c2eeaaab21f3abdde1cf06a21d63eb691fd1308af00ede883d0229"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.671459 4719 generic.go:334] "Generic (PLEG): container finished" podID="fdc568d9-e75c-4f9b-a426-4dbc175a9f44" containerID="2c9715a195700a9e7cd49c8aea13f28966c101cf2b45bc29b3a0470cb83f8e53" exitCode=0 Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.671535 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d2d3-account-create-update-8lpdk" event={"ID":"fdc568d9-e75c-4f9b-a426-4dbc175a9f44","Type":"ContainerDied","Data":"2c9715a195700a9e7cd49c8aea13f28966c101cf2b45bc29b3a0470cb83f8e53"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.678743 4719 generic.go:334] "Generic (PLEG): container finished" podID="c0e6d128-ab9b-4a15-bef7-c986942dc16f" containerID="530d8a4b0fa480da3faa6731792800736b86d48a1107001a800489dc05bafb7b" exitCode=0 Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.678797 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6f1e-account-create-update-bps6c" event={"ID":"c0e6d128-ab9b-4a15-bef7-c986942dc16f","Type":"ContainerDied","Data":"530d8a4b0fa480da3faa6731792800736b86d48a1107001a800489dc05bafb7b"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.681182 4719 generic.go:334] "Generic (PLEG): container finished" podID="83819057-000b-4f9e-b8fc-3478bc403f4e" containerID="fb301aeb8662ca386a1b946e9d89d0d5126b0c82f9f845b0c7de32fa368ebbca" exitCode=0 Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.681242 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-539a-account-create-update-zqll7" event={"ID":"83819057-000b-4f9e-b8fc-3478bc403f4e","Type":"ContainerDied","Data":"fb301aeb8662ca386a1b946e9d89d0d5126b0c82f9f845b0c7de32fa368ebbca"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.686588 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tcmf2" event={"ID":"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752","Type":"ContainerStarted","Data":"293b2bef238e11454bddbe8d8d1e8bfab6dea3f513dcaac1bfe8995a320a3589"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.686615 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tcmf2" event={"ID":"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752","Type":"ContainerStarted","Data":"dda78f6108a4f050fc980a3b9f164fbb77f4aeb705a8f5d08cb4d6b043d90b55"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.690719 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-75qg9" podStartSLOduration=3.6907083 podStartE2EDuration="3.6907083s" podCreationTimestamp="2025-12-15 12:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:21.688828292 +0000 UTC m=+1202.631121322" watchObservedRunningTime="2025-12-15 12:37:21.6907083 +0000 UTC m=+1202.633001330" Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.698451 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c6wxw" event={"ID":"4c64d8b6-1005-484e-b068-95d462167a8c","Type":"ContainerStarted","Data":"81f11fd8a4da84febea345ef05cbb09baedb39824bc2f7854d829b29ee8224fa"} Dec 15 12:37:21 crc kubenswrapper[4719]: I1215 12:37:21.733401 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-tcmf2" podStartSLOduration=3.733385357 podStartE2EDuration="3.733385357s" podCreationTimestamp="2025-12-15 12:37:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:21.710280238 +0000 UTC m=+1202.652573268" watchObservedRunningTime="2025-12-15 12:37:21.733385357 +0000 UTC m=+1202.675678387" Dec 15 12:37:22 crc kubenswrapper[4719]: I1215 12:37:22.707766 4719 generic.go:334] "Generic (PLEG): container finished" podID="d8d04b79-2fb3-44a2-9b61-d6ab75ad9752" containerID="293b2bef238e11454bddbe8d8d1e8bfab6dea3f513dcaac1bfe8995a320a3589" exitCode=0 Dec 15 12:37:22 crc kubenswrapper[4719]: I1215 12:37:22.708009 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tcmf2" event={"ID":"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752","Type":"ContainerDied","Data":"293b2bef238e11454bddbe8d8d1e8bfab6dea3f513dcaac1bfe8995a320a3589"} Dec 15 12:37:22 crc kubenswrapper[4719]: I1215 12:37:22.717595 4719 generic.go:334] "Generic (PLEG): container finished" podID="0381936c-f528-4662-8a86-558c341065c0" containerID="5f094fe3136a3f3ce0225bc95beac5be77d50f6bddf8229c8d1be2f03cab6b51" exitCode=0 Dec 15 12:37:22 crc kubenswrapper[4719]: I1215 12:37:22.717874 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-75qg9" event={"ID":"0381936c-f528-4662-8a86-558c341065c0","Type":"ContainerDied","Data":"5f094fe3136a3f3ce0225bc95beac5be77d50f6bddf8229c8d1be2f03cab6b51"} Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.269339 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d2d3-account-create-update-8lpdk" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.300027 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-operator-scripts\") pod \"fdc568d9-e75c-4f9b-a426-4dbc175a9f44\" (UID: \"fdc568d9-e75c-4f9b-a426-4dbc175a9f44\") " Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.300069 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqv7s\" (UniqueName: \"kubernetes.io/projected/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-kube-api-access-cqv7s\") pod \"fdc568d9-e75c-4f9b-a426-4dbc175a9f44\" (UID: \"fdc568d9-e75c-4f9b-a426-4dbc175a9f44\") " Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.303332 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fdc568d9-e75c-4f9b-a426-4dbc175a9f44" (UID: "fdc568d9-e75c-4f9b-a426-4dbc175a9f44"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.307389 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-kube-api-access-cqv7s" (OuterVolumeSpecName: "kube-api-access-cqv7s") pod "fdc568d9-e75c-4f9b-a426-4dbc175a9f44" (UID: "fdc568d9-e75c-4f9b-a426-4dbc175a9f44"). InnerVolumeSpecName "kube-api-access-cqv7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.386000 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.392663 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-539a-account-create-update-zqll7" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.400749 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lgf62" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.401570 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.401604 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqv7s\" (UniqueName: \"kubernetes.io/projected/fdc568d9-e75c-4f9b-a426-4dbc175a9f44-kube-api-access-cqv7s\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.431770 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6f1e-account-create-update-bps6c" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.472705 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-g4d2t"] Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.473067 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-g4d2t" podUID="dda77ba2-3b59-4946-940b-8c8987f2884b" containerName="dnsmasq-dns" containerID="cri-o://1e9f71ff98b60da5c1ac92b2eb5821eba7caa3d3a4c4b54b042ea00689040490" gracePeriod=10 Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.502310 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0e6d128-ab9b-4a15-bef7-c986942dc16f-operator-scripts\") pod \"c0e6d128-ab9b-4a15-bef7-c986942dc16f\" (UID: \"c0e6d128-ab9b-4a15-bef7-c986942dc16f\") " Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.502387 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj725\" (UniqueName: \"kubernetes.io/projected/83819057-000b-4f9e-b8fc-3478bc403f4e-kube-api-access-wj725\") pod \"83819057-000b-4f9e-b8fc-3478bc403f4e\" (UID: \"83819057-000b-4f9e-b8fc-3478bc403f4e\") " Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.502440 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz54b\" (UniqueName: \"kubernetes.io/projected/c0e6d128-ab9b-4a15-bef7-c986942dc16f-kube-api-access-vz54b\") pod \"c0e6d128-ab9b-4a15-bef7-c986942dc16f\" (UID: \"c0e6d128-ab9b-4a15-bef7-c986942dc16f\") " Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.502477 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-operator-scripts\") pod \"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1\" (UID: \"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1\") " Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.502554 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8kpw\" (UniqueName: \"kubernetes.io/projected/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-kube-api-access-r8kpw\") pod \"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1\" (UID: \"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1\") " Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.502611 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83819057-000b-4f9e-b8fc-3478bc403f4e-operator-scripts\") pod \"83819057-000b-4f9e-b8fc-3478bc403f4e\" (UID: \"83819057-000b-4f9e-b8fc-3478bc403f4e\") " Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.503978 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0e6d128-ab9b-4a15-bef7-c986942dc16f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c0e6d128-ab9b-4a15-bef7-c986942dc16f" (UID: "c0e6d128-ab9b-4a15-bef7-c986942dc16f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.507783 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e2ec6917-fd83-4b5b-b904-6a0e4d453cf1" (UID: "e2ec6917-fd83-4b5b-b904-6a0e4d453cf1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.508287 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83819057-000b-4f9e-b8fc-3478bc403f4e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83819057-000b-4f9e-b8fc-3478bc403f4e" (UID: "83819057-000b-4f9e-b8fc-3478bc403f4e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.524613 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0e6d128-ab9b-4a15-bef7-c986942dc16f-kube-api-access-vz54b" (OuterVolumeSpecName: "kube-api-access-vz54b") pod "c0e6d128-ab9b-4a15-bef7-c986942dc16f" (UID: "c0e6d128-ab9b-4a15-bef7-c986942dc16f"). InnerVolumeSpecName "kube-api-access-vz54b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.528173 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-kube-api-access-r8kpw" (OuterVolumeSpecName: "kube-api-access-r8kpw") pod "e2ec6917-fd83-4b5b-b904-6a0e4d453cf1" (UID: "e2ec6917-fd83-4b5b-b904-6a0e4d453cf1"). InnerVolumeSpecName "kube-api-access-r8kpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.531183 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83819057-000b-4f9e-b8fc-3478bc403f4e-kube-api-access-wj725" (OuterVolumeSpecName: "kube-api-access-wj725") pod "83819057-000b-4f9e-b8fc-3478bc403f4e" (UID: "83819057-000b-4f9e-b8fc-3478bc403f4e"). InnerVolumeSpecName "kube-api-access-wj725". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.605017 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj725\" (UniqueName: \"kubernetes.io/projected/83819057-000b-4f9e-b8fc-3478bc403f4e-kube-api-access-wj725\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.605060 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz54b\" (UniqueName: \"kubernetes.io/projected/c0e6d128-ab9b-4a15-bef7-c986942dc16f-kube-api-access-vz54b\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.605070 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.605080 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8kpw\" (UniqueName: \"kubernetes.io/projected/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1-kube-api-access-r8kpw\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.605088 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83819057-000b-4f9e-b8fc-3478bc403f4e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.605097 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0e6d128-ab9b-4a15-bef7-c986942dc16f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.731155 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-lgf62" event={"ID":"e2ec6917-fd83-4b5b-b904-6a0e4d453cf1","Type":"ContainerDied","Data":"c9ed5081a49c8d75ec7de04fa383fefa1599366d248181b2be27eab54b5f12e3"} Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.731221 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9ed5081a49c8d75ec7de04fa383fefa1599366d248181b2be27eab54b5f12e3" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.731165 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-lgf62" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.735499 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d2d3-account-create-update-8lpdk" event={"ID":"fdc568d9-e75c-4f9b-a426-4dbc175a9f44","Type":"ContainerDied","Data":"554b9096284f62436d80500e92bec603f32d5e1664fc475fc25aaae3ab60aca9"} Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.735544 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="554b9096284f62436d80500e92bec603f32d5e1664fc475fc25aaae3ab60aca9" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.735610 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d2d3-account-create-update-8lpdk" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.741679 4719 generic.go:334] "Generic (PLEG): container finished" podID="dda77ba2-3b59-4946-940b-8c8987f2884b" containerID="1e9f71ff98b60da5c1ac92b2eb5821eba7caa3d3a4c4b54b042ea00689040490" exitCode=0 Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.741786 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-g4d2t" event={"ID":"dda77ba2-3b59-4946-940b-8c8987f2884b","Type":"ContainerDied","Data":"1e9f71ff98b60da5c1ac92b2eb5821eba7caa3d3a4c4b54b042ea00689040490"} Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.745073 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6f1e-account-create-update-bps6c" event={"ID":"c0e6d128-ab9b-4a15-bef7-c986942dc16f","Type":"ContainerDied","Data":"25ec6464b837bfff94aaef1302e567575f14dc77711f9f028071fcfe941a1e61"} Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.745113 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25ec6464b837bfff94aaef1302e567575f14dc77711f9f028071fcfe941a1e61" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.745160 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6f1e-account-create-update-bps6c" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.755533 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-539a-account-create-update-zqll7" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.755832 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-539a-account-create-update-zqll7" event={"ID":"83819057-000b-4f9e-b8fc-3478bc403f4e","Type":"ContainerDied","Data":"0064de5b30526464acc268f4f5bcc296566bac160d7527c71a1c266bea08577c"} Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.755900 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0064de5b30526464acc268f4f5bcc296566bac160d7527c71a1c266bea08577c" Dec 15 12:37:23 crc kubenswrapper[4719]: I1215 12:37:23.774899 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"e658059578b8db5d03eadd14b3655cfd92d4f976a054850efce35b9b36590b02"} Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.071579 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.130602 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-nb\") pod \"dda77ba2-3b59-4946-940b-8c8987f2884b\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.130664 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-dns-svc\") pod \"dda77ba2-3b59-4946-940b-8c8987f2884b\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.130750 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-config\") pod \"dda77ba2-3b59-4946-940b-8c8987f2884b\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.130887 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-sb\") pod \"dda77ba2-3b59-4946-940b-8c8987f2884b\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.130954 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9vw4\" (UniqueName: \"kubernetes.io/projected/dda77ba2-3b59-4946-940b-8c8987f2884b-kube-api-access-b9vw4\") pod \"dda77ba2-3b59-4946-940b-8c8987f2884b\" (UID: \"dda77ba2-3b59-4946-940b-8c8987f2884b\") " Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.135656 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dda77ba2-3b59-4946-940b-8c8987f2884b-kube-api-access-b9vw4" (OuterVolumeSpecName: "kube-api-access-b9vw4") pod "dda77ba2-3b59-4946-940b-8c8987f2884b" (UID: "dda77ba2-3b59-4946-940b-8c8987f2884b"). InnerVolumeSpecName "kube-api-access-b9vw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.232877 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9vw4\" (UniqueName: \"kubernetes.io/projected/dda77ba2-3b59-4946-940b-8c8987f2884b-kube-api-access-b9vw4\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.256084 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dda77ba2-3b59-4946-940b-8c8987f2884b" (UID: "dda77ba2-3b59-4946-940b-8c8987f2884b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.282249 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-config" (OuterVolumeSpecName: "config") pod "dda77ba2-3b59-4946-940b-8c8987f2884b" (UID: "dda77ba2-3b59-4946-940b-8c8987f2884b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.283479 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dda77ba2-3b59-4946-940b-8c8987f2884b" (UID: "dda77ba2-3b59-4946-940b-8c8987f2884b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.297729 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dda77ba2-3b59-4946-940b-8c8987f2884b" (UID: "dda77ba2-3b59-4946-940b-8c8987f2884b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.340771 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.340802 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.340813 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.340822 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dda77ba2-3b59-4946-940b-8c8987f2884b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.420658 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-75qg9" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.424512 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tcmf2" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.543179 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp6wt\" (UniqueName: \"kubernetes.io/projected/0381936c-f528-4662-8a86-558c341065c0-kube-api-access-hp6wt\") pod \"0381936c-f528-4662-8a86-558c341065c0\" (UID: \"0381936c-f528-4662-8a86-558c341065c0\") " Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.543263 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-operator-scripts\") pod \"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752\" (UID: \"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752\") " Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.543319 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf92m\" (UniqueName: \"kubernetes.io/projected/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-kube-api-access-zf92m\") pod \"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752\" (UID: \"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752\") " Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.543408 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0381936c-f528-4662-8a86-558c341065c0-operator-scripts\") pod \"0381936c-f528-4662-8a86-558c341065c0\" (UID: \"0381936c-f528-4662-8a86-558c341065c0\") " Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.547657 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0381936c-f528-4662-8a86-558c341065c0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0381936c-f528-4662-8a86-558c341065c0" (UID: "0381936c-f528-4662-8a86-558c341065c0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.552204 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-kube-api-access-zf92m" (OuterVolumeSpecName: "kube-api-access-zf92m") pod "d8d04b79-2fb3-44a2-9b61-d6ab75ad9752" (UID: "d8d04b79-2fb3-44a2-9b61-d6ab75ad9752"). InnerVolumeSpecName "kube-api-access-zf92m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.552407 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d8d04b79-2fb3-44a2-9b61-d6ab75ad9752" (UID: "d8d04b79-2fb3-44a2-9b61-d6ab75ad9752"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.557761 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0381936c-f528-4662-8a86-558c341065c0-kube-api-access-hp6wt" (OuterVolumeSpecName: "kube-api-access-hp6wt") pod "0381936c-f528-4662-8a86-558c341065c0" (UID: "0381936c-f528-4662-8a86-558c341065c0"). InnerVolumeSpecName "kube-api-access-hp6wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.645251 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0381936c-f528-4662-8a86-558c341065c0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.645284 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp6wt\" (UniqueName: \"kubernetes.io/projected/0381936c-f528-4662-8a86-558c341065c0-kube-api-access-hp6wt\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.645293 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.645304 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf92m\" (UniqueName: \"kubernetes.io/projected/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752-kube-api-access-zf92m\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.786096 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-75qg9" event={"ID":"0381936c-f528-4662-8a86-558c341065c0","Type":"ContainerDied","Data":"c91bf3e196c2eeaaab21f3abdde1cf06a21d63eb691fd1308af00ede883d0229"} Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.786147 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c91bf3e196c2eeaaab21f3abdde1cf06a21d63eb691fd1308af00ede883d0229" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.786465 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-75qg9" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.792915 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-g4d2t" event={"ID":"dda77ba2-3b59-4946-940b-8c8987f2884b","Type":"ContainerDied","Data":"379e77f8161bcdb483565f2d1ef2c586af1cda5db00b990b6150d2f18e1906d4"} Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.792976 4719 scope.go:117] "RemoveContainer" containerID="1e9f71ff98b60da5c1ac92b2eb5821eba7caa3d3a4c4b54b042ea00689040490" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.792978 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-g4d2t" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.796157 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tcmf2" event={"ID":"d8d04b79-2fb3-44a2-9b61-d6ab75ad9752","Type":"ContainerDied","Data":"dda78f6108a4f050fc980a3b9f164fbb77f4aeb705a8f5d08cb4d6b043d90b55"} Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.796200 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dda78f6108a4f050fc980a3b9f164fbb77f4aeb705a8f5d08cb4d6b043d90b55" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.796306 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tcmf2" Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.823054 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"434390b1173237fdfdfc9173351fa1b46826ce18828ab0339fb8caa22daa2742"} Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.823096 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"ab13121d1d7c1cefd17d6399d445afbf54d595bb903f9162871f54670631e168"} Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.823112 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"fc9a608b00d0dd8c6ddbec71120dfbc1eef49ea082db40541f26a35c5b21b17d"} Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.845832 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-g4d2t"] Dec 15 12:37:24 crc kubenswrapper[4719]: I1215 12:37:24.852655 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-g4d2t"] Dec 15 12:37:25 crc kubenswrapper[4719]: I1215 12:37:25.499980 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dda77ba2-3b59-4946-940b-8c8987f2884b" path="/var/lib/kubelet/pods/dda77ba2-3b59-4946-940b-8c8987f2884b/volumes" Dec 15 12:37:27 crc kubenswrapper[4719]: I1215 12:37:27.506925 4719 scope.go:117] "RemoveContainer" containerID="ba4129088dbebcd0ed5673846e7fe28747d58d295ddfca0b1402ed057144f27e" Dec 15 12:37:27 crc kubenswrapper[4719]: I1215 12:37:27.849843 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c6wxw" event={"ID":"4c64d8b6-1005-484e-b068-95d462167a8c","Type":"ContainerStarted","Data":"d8fc87b95d5a2eaed35562d83d92e0bdefc4ed210c7b6338f3037671ab1736f6"} Dec 15 12:37:28 crc kubenswrapper[4719]: I1215 12:37:28.802564 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-g4d2t" podUID="dda77ba2-3b59-4946-940b-8c8987f2884b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.120:5353: i/o timeout" Dec 15 12:37:28 crc kubenswrapper[4719]: I1215 12:37:28.867372 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"ae53db9d9193da6608a7093eb9200e8b5f5ae1c3a699a72ef8386243b5bc1ad5"} Dec 15 12:37:28 crc kubenswrapper[4719]: I1215 12:37:28.867427 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"99f900c15ee10650ccf25278fb34b4f32fb65b256df47ebfe50e8a2c51a7a27c"} Dec 15 12:37:29 crc kubenswrapper[4719]: I1215 12:37:29.880804 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"91da51d66bd6b7b7bee8c80eb5916f25b6dd419fa97682981b2453be30fc1c8d"} Dec 15 12:37:30 crc kubenswrapper[4719]: I1215 12:37:30.901811 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"797eec3202731016c9c34cca3ef9bd605d1fe88c391b13d781aa4e6def623272"} Dec 15 12:37:30 crc kubenswrapper[4719]: I1215 12:37:30.902536 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"4b791e658957f69852ccbdb1e4de7661bd0af11cab13266743fcf0cb17be7e8f"} Dec 15 12:37:30 crc kubenswrapper[4719]: I1215 12:37:30.902602 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"5bf83fdbfc223b5a79b71f18662325c4393f9cc1372f2ef3df32733c4d7ba963"} Dec 15 12:37:31 crc kubenswrapper[4719]: I1215 12:37:31.916450 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"651edea8-3300-463c-a0fa-abd75f6056e0","Type":"ContainerStarted","Data":"c14e4ce14e19459d1ba8c15ca4f662abdb2a19c9737081d07db90d42a64f21c8"} Dec 15 12:37:31 crc kubenswrapper[4719]: I1215 12:37:31.964298 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=38.111946621 podStartE2EDuration="48.964283386s" podCreationTimestamp="2025-12-15 12:36:43 +0000 UTC" firstStartedPulling="2025-12-15 12:37:17.514288476 +0000 UTC m=+1198.456581506" lastFinishedPulling="2025-12-15 12:37:28.366625241 +0000 UTC m=+1209.308918271" observedRunningTime="2025-12-15 12:37:31.962484221 +0000 UTC m=+1212.904777241" watchObservedRunningTime="2025-12-15 12:37:31.964283386 +0000 UTC m=+1212.906576416" Dec 15 12:37:31 crc kubenswrapper[4719]: I1215 12:37:31.970774 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-c6wxw" podStartSLOduration=7.095372014 podStartE2EDuration="13.970756348s" podCreationTimestamp="2025-12-15 12:37:18 +0000 UTC" firstStartedPulling="2025-12-15 12:37:20.691935531 +0000 UTC m=+1201.634228561" lastFinishedPulling="2025-12-15 12:37:27.567319855 +0000 UTC m=+1208.509612895" observedRunningTime="2025-12-15 12:37:27.880723071 +0000 UTC m=+1208.823016101" watchObservedRunningTime="2025-12-15 12:37:31.970756348 +0000 UTC m=+1212.913049378" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.255642 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-xcjn4"] Dec 15 12:37:32 crc kubenswrapper[4719]: E1215 12:37:32.256051 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d04b79-2fb3-44a2-9b61-d6ab75ad9752" containerName="mariadb-database-create" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256073 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d04b79-2fb3-44a2-9b61-d6ab75ad9752" containerName="mariadb-database-create" Dec 15 12:37:32 crc kubenswrapper[4719]: E1215 12:37:32.256087 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83819057-000b-4f9e-b8fc-3478bc403f4e" containerName="mariadb-account-create-update" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256095 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="83819057-000b-4f9e-b8fc-3478bc403f4e" containerName="mariadb-account-create-update" Dec 15 12:37:32 crc kubenswrapper[4719]: E1215 12:37:32.256122 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2ec6917-fd83-4b5b-b904-6a0e4d453cf1" containerName="mariadb-database-create" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256131 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2ec6917-fd83-4b5b-b904-6a0e4d453cf1" containerName="mariadb-database-create" Dec 15 12:37:32 crc kubenswrapper[4719]: E1215 12:37:32.256148 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dda77ba2-3b59-4946-940b-8c8987f2884b" containerName="dnsmasq-dns" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256156 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="dda77ba2-3b59-4946-940b-8c8987f2884b" containerName="dnsmasq-dns" Dec 15 12:37:32 crc kubenswrapper[4719]: E1215 12:37:32.256171 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0381936c-f528-4662-8a86-558c341065c0" containerName="mariadb-database-create" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256178 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="0381936c-f528-4662-8a86-558c341065c0" containerName="mariadb-database-create" Dec 15 12:37:32 crc kubenswrapper[4719]: E1215 12:37:32.256188 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dda77ba2-3b59-4946-940b-8c8987f2884b" containerName="init" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256197 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="dda77ba2-3b59-4946-940b-8c8987f2884b" containerName="init" Dec 15 12:37:32 crc kubenswrapper[4719]: E1215 12:37:32.256209 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0e6d128-ab9b-4a15-bef7-c986942dc16f" containerName="mariadb-account-create-update" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256217 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0e6d128-ab9b-4a15-bef7-c986942dc16f" containerName="mariadb-account-create-update" Dec 15 12:37:32 crc kubenswrapper[4719]: E1215 12:37:32.256233 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdc568d9-e75c-4f9b-a426-4dbc175a9f44" containerName="mariadb-account-create-update" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256241 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdc568d9-e75c-4f9b-a426-4dbc175a9f44" containerName="mariadb-account-create-update" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256454 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2ec6917-fd83-4b5b-b904-6a0e4d453cf1" containerName="mariadb-database-create" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256477 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d04b79-2fb3-44a2-9b61-d6ab75ad9752" containerName="mariadb-database-create" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256487 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="dda77ba2-3b59-4946-940b-8c8987f2884b" containerName="dnsmasq-dns" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256499 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="83819057-000b-4f9e-b8fc-3478bc403f4e" containerName="mariadb-account-create-update" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256512 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdc568d9-e75c-4f9b-a426-4dbc175a9f44" containerName="mariadb-account-create-update" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256525 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0e6d128-ab9b-4a15-bef7-c986942dc16f" containerName="mariadb-account-create-update" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.256539 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="0381936c-f528-4662-8a86-558c341065c0" containerName="mariadb-database-create" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.257422 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.259471 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.279009 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-xcjn4"] Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.297790 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.297866 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.298023 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffh4r\" (UniqueName: \"kubernetes.io/projected/7bba374b-c278-4218-917b-906554fffacd-kube-api-access-ffh4r\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.298148 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.298292 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.298376 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-config\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.400222 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.400291 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-config\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.400347 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.400386 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.400421 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffh4r\" (UniqueName: \"kubernetes.io/projected/7bba374b-c278-4218-917b-906554fffacd-kube-api-access-ffh4r\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.400473 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.401235 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-config\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.401312 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.401383 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.401404 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.401386 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.428220 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffh4r\" (UniqueName: \"kubernetes.io/projected/7bba374b-c278-4218-917b-906554fffacd-kube-api-access-ffh4r\") pod \"dnsmasq-dns-7ff5475cc9-xcjn4\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:32 crc kubenswrapper[4719]: I1215 12:37:32.573162 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:33 crc kubenswrapper[4719]: I1215 12:37:33.080956 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-xcjn4"] Dec 15 12:37:33 crc kubenswrapper[4719]: I1215 12:37:33.939376 4719 generic.go:334] "Generic (PLEG): container finished" podID="7bba374b-c278-4218-917b-906554fffacd" containerID="b1f1eaba2fa3f346a2dc54a03925118f0d23158e15d94315413446b28826ec8b" exitCode=0 Dec 15 12:37:33 crc kubenswrapper[4719]: I1215 12:37:33.939423 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" event={"ID":"7bba374b-c278-4218-917b-906554fffacd","Type":"ContainerDied","Data":"b1f1eaba2fa3f346a2dc54a03925118f0d23158e15d94315413446b28826ec8b"} Dec 15 12:37:33 crc kubenswrapper[4719]: I1215 12:37:33.939743 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" event={"ID":"7bba374b-c278-4218-917b-906554fffacd","Type":"ContainerStarted","Data":"3c0fdb040cc6fcd32e64fc32914669eea738f194ca0a62220ee9256b40279c59"} Dec 15 12:37:34 crc kubenswrapper[4719]: I1215 12:37:34.950238 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" event={"ID":"7bba374b-c278-4218-917b-906554fffacd","Type":"ContainerStarted","Data":"d96f4a08c4fc8b11192c93a930e3bfd07954f6026d9c01b8eb64232b220f0eb9"} Dec 15 12:37:34 crc kubenswrapper[4719]: I1215 12:37:34.950556 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:34 crc kubenswrapper[4719]: I1215 12:37:34.953092 4719 generic.go:334] "Generic (PLEG): container finished" podID="4c64d8b6-1005-484e-b068-95d462167a8c" containerID="d8fc87b95d5a2eaed35562d83d92e0bdefc4ed210c7b6338f3037671ab1736f6" exitCode=0 Dec 15 12:37:34 crc kubenswrapper[4719]: I1215 12:37:34.953146 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c6wxw" event={"ID":"4c64d8b6-1005-484e-b068-95d462167a8c","Type":"ContainerDied","Data":"d8fc87b95d5a2eaed35562d83d92e0bdefc4ed210c7b6338f3037671ab1736f6"} Dec 15 12:37:34 crc kubenswrapper[4719]: I1215 12:37:34.973238 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" podStartSLOduration=2.973216295 podStartE2EDuration="2.973216295s" podCreationTimestamp="2025-12-15 12:37:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:34.972488753 +0000 UTC m=+1215.914781783" watchObservedRunningTime="2025-12-15 12:37:34.973216295 +0000 UTC m=+1215.915509325" Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.291684 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.364501 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-config-data\") pod \"4c64d8b6-1005-484e-b068-95d462167a8c\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.364658 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k994q\" (UniqueName: \"kubernetes.io/projected/4c64d8b6-1005-484e-b068-95d462167a8c-kube-api-access-k994q\") pod \"4c64d8b6-1005-484e-b068-95d462167a8c\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.364702 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-combined-ca-bundle\") pod \"4c64d8b6-1005-484e-b068-95d462167a8c\" (UID: \"4c64d8b6-1005-484e-b068-95d462167a8c\") " Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.369916 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c64d8b6-1005-484e-b068-95d462167a8c-kube-api-access-k994q" (OuterVolumeSpecName: "kube-api-access-k994q") pod "4c64d8b6-1005-484e-b068-95d462167a8c" (UID: "4c64d8b6-1005-484e-b068-95d462167a8c"). InnerVolumeSpecName "kube-api-access-k994q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.395276 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c64d8b6-1005-484e-b068-95d462167a8c" (UID: "4c64d8b6-1005-484e-b068-95d462167a8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.407670 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-config-data" (OuterVolumeSpecName: "config-data") pod "4c64d8b6-1005-484e-b068-95d462167a8c" (UID: "4c64d8b6-1005-484e-b068-95d462167a8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.467005 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k994q\" (UniqueName: \"kubernetes.io/projected/4c64d8b6-1005-484e-b068-95d462167a8c-kube-api-access-k994q\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.467053 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.467102 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c64d8b6-1005-484e-b068-95d462167a8c-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.975209 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c6wxw" event={"ID":"4c64d8b6-1005-484e-b068-95d462167a8c","Type":"ContainerDied","Data":"81f11fd8a4da84febea345ef05cbb09baedb39824bc2f7854d829b29ee8224fa"} Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.975793 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81f11fd8a4da84febea345ef05cbb09baedb39824bc2f7854d829b29ee8224fa" Dec 15 12:37:36 crc kubenswrapper[4719]: I1215 12:37:36.975300 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c6wxw" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.190306 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jgtsh"] Dec 15 12:37:37 crc kubenswrapper[4719]: E1215 12:37:37.190948 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c64d8b6-1005-484e-b068-95d462167a8c" containerName="keystone-db-sync" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.191020 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c64d8b6-1005-484e-b068-95d462167a8c" containerName="keystone-db-sync" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.191233 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c64d8b6-1005-484e-b068-95d462167a8c" containerName="keystone-db-sync" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.191992 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.196079 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kdx8k" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.196294 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.196420 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.199847 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.201399 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.209004 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jgtsh"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.280581 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-credential-keys\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.280630 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-scripts\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.280678 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-fernet-keys\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.280744 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzz67\" (UniqueName: \"kubernetes.io/projected/9d518023-a040-4289-b78b-163624c30b2a-kube-api-access-zzz67\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.280844 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-config-data\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.280924 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-combined-ca-bundle\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.317070 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-xcjn4"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.317271 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" podUID="7bba374b-c278-4218-917b-906554fffacd" containerName="dnsmasq-dns" containerID="cri-o://d96f4a08c4fc8b11192c93a930e3bfd07954f6026d9c01b8eb64232b220f0eb9" gracePeriod=10 Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.360542 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.361787 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.381937 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-credential-keys\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.381988 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-scripts\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.382016 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.382064 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-fernet-keys\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.382080 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.382106 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzz67\" (UniqueName: \"kubernetes.io/projected/9d518023-a040-4289-b78b-163624c30b2a-kube-api-access-zzz67\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.382122 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.382153 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-config-data\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.382185 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.382216 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwhdl\" (UniqueName: \"kubernetes.io/projected/8370e556-0bec-4693-8440-644163299e5a-kube-api-access-gwhdl\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.382244 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-combined-ca-bundle\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.382262 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-config\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.387580 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-credential-keys\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.393670 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-scripts\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.404214 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-config-data\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.408131 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-combined-ca-bundle\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.410457 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-fernet-keys\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.424379 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.462811 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzz67\" (UniqueName: \"kubernetes.io/projected/9d518023-a040-4289-b78b-163624c30b2a-kube-api-access-zzz67\") pod \"keystone-bootstrap-jgtsh\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.488797 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.488872 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwhdl\" (UniqueName: \"kubernetes.io/projected/8370e556-0bec-4693-8440-644163299e5a-kube-api-access-gwhdl\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.488908 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-config\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.488938 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.488983 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.489005 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.489870 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.490123 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-config\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.490415 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.490572 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.493331 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.512419 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.547044 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwhdl\" (UniqueName: \"kubernetes.io/projected/8370e556-0bec-4693-8440-644163299e5a-kube-api-access-gwhdl\") pod \"dnsmasq-dns-5c5cc7c5ff-8r8vl\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.591228 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5dc4d6cbb7-scwpt"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.592962 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.619570 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.619900 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.620016 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-xhzkv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.620049 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.650662 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5dc4d6cbb7-scwpt"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.678964 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.679108 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-gprbv"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.680276 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.685275 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.685629 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.686014 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dwxb2" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.700558 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fbpq\" (UniqueName: \"kubernetes.io/projected/27a897af-5729-4d7a-94a6-8318aa88eff5-kube-api-access-7fbpq\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.700615 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-config-data\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.700645 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27a897af-5729-4d7a-94a6-8318aa88eff5-horizon-secret-key\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.700694 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a897af-5729-4d7a-94a6-8318aa88eff5-logs\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.700711 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-scripts\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.747947 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gprbv"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.802649 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-scripts\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.803308 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-etc-machine-id\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.803424 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-combined-ca-bundle\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.803495 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a897af-5729-4d7a-94a6-8318aa88eff5-logs\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.803560 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-scripts\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.803674 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-config-data\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.803748 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4drnb\" (UniqueName: \"kubernetes.io/projected/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-kube-api-access-4drnb\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.803843 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fbpq\" (UniqueName: \"kubernetes.io/projected/27a897af-5729-4d7a-94a6-8318aa88eff5-kube-api-access-7fbpq\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.803940 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-config-data\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.804002 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-db-sync-config-data\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.804098 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27a897af-5729-4d7a-94a6-8318aa88eff5-horizon-secret-key\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.804639 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a897af-5729-4d7a-94a6-8318aa88eff5-logs\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.805096 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-scripts\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.805614 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-config-data\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.817869 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.819463 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27a897af-5729-4d7a-94a6-8318aa88eff5-horizon-secret-key\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.828311 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.866574 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.866847 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.875063 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fbpq\" (UniqueName: \"kubernetes.io/projected/27a897af-5729-4d7a-94a6-8318aa88eff5-kube-api-access-7fbpq\") pod \"horizon-5dc4d6cbb7-scwpt\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.876080 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.908307 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-etc-machine-id\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.908359 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-combined-ca-bundle\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.908405 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-etc-machine-id\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909123 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909171 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-config-data\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909197 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-config-data\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909228 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4drnb\" (UniqueName: \"kubernetes.io/projected/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-kube-api-access-4drnb\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909251 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-scripts\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909301 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-db-sync-config-data\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909330 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909351 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-log-httpd\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909367 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmmgs\" (UniqueName: \"kubernetes.io/projected/edd905bf-6348-446d-abe8-fb632760bc1f-kube-api-access-rmmgs\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909387 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-run-httpd\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.909411 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-scripts\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.926774 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-config-data\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.937843 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.946172 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-9vwxh"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.949710 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.966505 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.966698 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.966877 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-g5z85" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.974623 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-combined-ca-bundle\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.977034 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9vwxh"] Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.981440 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-db-sync-config-data\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.983548 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-scripts\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:37 crc kubenswrapper[4719]: I1215 12:37:37.995369 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4drnb\" (UniqueName: \"kubernetes.io/projected/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-kube-api-access-4drnb\") pod \"cinder-db-sync-gprbv\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.016973 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-8qr2l"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.017993 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.019267 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-combined-ca-bundle\") pod \"neutron-db-sync-9vwxh\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.019301 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.019324 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-log-httpd\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.019339 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmmgs\" (UniqueName: \"kubernetes.io/projected/edd905bf-6348-446d-abe8-fb632760bc1f-kube-api-access-rmmgs\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.019360 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-run-httpd\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.019381 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c89dt\" (UniqueName: \"kubernetes.io/projected/3243ab84-0a4c-4c74-83de-693c1c96e89e-kube-api-access-c89dt\") pod \"neutron-db-sync-9vwxh\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.019426 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-config\") pod \"neutron-db-sync-9vwxh\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.019450 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.019476 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-config-data\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.019506 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-scripts\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.020373 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-run-httpd\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.021466 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-log-httpd\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.022947 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-scripts\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.086551 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.088354 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.089309 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-4dhqg" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.089894 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-config-data\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.090619 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.090697 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-8qr2l"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.107601 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gprbv" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.180388 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmmgs\" (UniqueName: \"kubernetes.io/projected/edd905bf-6348-446d-abe8-fb632760bc1f-kube-api-access-rmmgs\") pod \"ceilometer-0\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.224233 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-combined-ca-bundle\") pod \"neutron-db-sync-9vwxh\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.224320 4719 generic.go:334] "Generic (PLEG): container finished" podID="7bba374b-c278-4218-917b-906554fffacd" containerID="d96f4a08c4fc8b11192c93a930e3bfd07954f6026d9c01b8eb64232b220f0eb9" exitCode=0 Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.224377 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" event={"ID":"7bba374b-c278-4218-917b-906554fffacd","Type":"ContainerDied","Data":"d96f4a08c4fc8b11192c93a930e3bfd07954f6026d9c01b8eb64232b220f0eb9"} Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.224368 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c89dt\" (UniqueName: \"kubernetes.io/projected/3243ab84-0a4c-4c74-83de-693c1c96e89e-kube-api-access-c89dt\") pod \"neutron-db-sync-9vwxh\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.224559 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-config\") pod \"neutron-db-sync-9vwxh\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.224641 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxzg5\" (UniqueName: \"kubernetes.io/projected/58fea5e6-9ff9-4a44-be8a-311735af3e4b-kube-api-access-fxzg5\") pod \"barbican-db-sync-8qr2l\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.224745 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-combined-ca-bundle\") pod \"barbican-db-sync-8qr2l\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.224828 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-db-sync-config-data\") pod \"barbican-db-sync-8qr2l\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.275869 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.277506 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nd852"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.279571 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.291710 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c89dt\" (UniqueName: \"kubernetes.io/projected/3243ab84-0a4c-4c74-83de-693c1c96e89e-kube-api-access-c89dt\") pod \"neutron-db-sync-9vwxh\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.295501 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.297521 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vbw9r" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.299124 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.317149 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-combined-ca-bundle\") pod \"neutron-db-sync-9vwxh\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.321537 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-config\") pod \"neutron-db-sync-9vwxh\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.324254 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nd852"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.333620 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f46b45545-rn64s"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.335190 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.381435 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f46b45545-rn64s"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.383657 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxzg5\" (UniqueName: \"kubernetes.io/projected/58fea5e6-9ff9-4a44-be8a-311735af3e4b-kube-api-access-fxzg5\") pod \"barbican-db-sync-8qr2l\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.383982 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-combined-ca-bundle\") pod \"barbican-db-sync-8qr2l\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.384078 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-db-sync-config-data\") pod \"barbican-db-sync-8qr2l\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.409462 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.410966 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.413565 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-db-sync-config-data\") pod \"barbican-db-sync-8qr2l\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.414485 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.431377 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-combined-ca-bundle\") pod \"barbican-db-sync-8qr2l\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.431520 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.431950 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.432063 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.432167 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-npxlb" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.486648 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv4vq\" (UniqueName: \"kubernetes.io/projected/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-kube-api-access-hv4vq\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.486712 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-scripts\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.486749 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-config-data\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.486810 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-logs\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.486830 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-scripts\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.486885 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-combined-ca-bundle\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.486912 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-config-data\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.486926 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlds7\" (UniqueName: \"kubernetes.io/projected/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-kube-api-access-rlds7\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.486950 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-horizon-secret-key\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.486966 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-logs\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.487752 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxzg5\" (UniqueName: \"kubernetes.io/projected/58fea5e6-9ff9-4a44-be8a-311735af3e4b-kube-api-access-fxzg5\") pod \"barbican-db-sync-8qr2l\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.523988 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.588121 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m64z\" (UniqueName: \"kubernetes.io/projected/ece6b2ac-3524-402c-a20b-bfc65dc046b4-kube-api-access-9m64z\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.588452 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-config-data\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.588495 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.588572 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-logs\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.588592 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.588614 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-scripts\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.588637 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-logs\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.588659 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.588687 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-combined-ca-bundle\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.590046 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-logs\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.594100 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.594177 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-config-data\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.594201 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlds7\" (UniqueName: \"kubernetes.io/projected/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-kube-api-access-rlds7\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.594248 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-horizon-secret-key\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.594273 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-logs\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.594320 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv4vq\" (UniqueName: \"kubernetes.io/projected/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-kube-api-access-hv4vq\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.594380 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-scripts\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.594399 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-scripts\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.594421 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-config-data\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.595694 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-config-data\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.599416 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-logs\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.600205 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-scripts\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.622334 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-scripts\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.623217 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-config-data\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.624936 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-combined-ca-bundle\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.661060 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.664229 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-horizon-secret-key\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.683378 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv4vq\" (UniqueName: \"kubernetes.io/projected/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-kube-api-access-hv4vq\") pod \"horizon-5f46b45545-rn64s\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.705990 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m64z\" (UniqueName: \"kubernetes.io/projected/ece6b2ac-3524-402c-a20b-bfc65dc046b4-kube-api-access-9m64z\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.706045 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.706103 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.706125 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-logs\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.706152 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.706191 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.706269 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-scripts\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.706290 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-config-data\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.708499 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-logs\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.710652 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.712192 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.731087 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-scripts\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.732337 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-config-data\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.741422 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.748531 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlds7\" (UniqueName: \"kubernetes.io/projected/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-kube-api-access-rlds7\") pod \"placement-db-sync-nd852\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " pod="openstack/placement-db-sync-nd852" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.748903 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.751801 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.782003 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.784495 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.786092 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.791311 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.800470 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.812357 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m64z\" (UniqueName: \"kubernetes.io/projected/ece6b2ac-3524-402c-a20b-bfc65dc046b4-kube-api-access-9m64z\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.826065 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.875945 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jgtsh"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.911290 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.935113 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.935173 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.935288 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.935394 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-logs\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.935452 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.935622 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pmzr\" (UniqueName: \"kubernetes.io/projected/5e309330-aa51-4178-a5cb-2f1ae1255255-kube-api-access-9pmzr\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.935730 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.935807 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.940026 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-52bzg"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.943904 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.970126 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-52bzg"] Dec 15 12:37:38 crc kubenswrapper[4719]: I1215 12:37:38.976779 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nd852" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.039198 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044150 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044205 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044275 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-config\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044388 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044421 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044491 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044548 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044586 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-logs\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044618 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044645 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044742 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pmzr\" (UniqueName: \"kubernetes.io/projected/5e309330-aa51-4178-a5cb-2f1ae1255255-kube-api-access-9pmzr\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044765 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.044786 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwmwv\" (UniqueName: \"kubernetes.io/projected/46738d80-d21e-43c5-8871-dba878a5986b-kube-api-access-bwmwv\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.046183 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-logs\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.046820 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.048845 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.059220 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.065115 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.067160 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.069220 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.084619 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl"] Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.094085 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.116274 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pmzr\" (UniqueName: \"kubernetes.io/projected/5e309330-aa51-4178-a5cb-2f1ae1255255-kube-api-access-9pmzr\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.142674 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.145800 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.145900 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.145923 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwmwv\" (UniqueName: \"kubernetes.io/projected/46738d80-d21e-43c5-8871-dba878a5986b-kube-api-access-bwmwv\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.145963 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.145993 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-config\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.146059 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.147528 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.149496 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.150116 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.150502 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.151261 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-config\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.183746 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwmwv\" (UniqueName: \"kubernetes.io/projected/46738d80-d21e-43c5-8871-dba878a5986b-kube-api-access-bwmwv\") pod \"dnsmasq-dns-8b5c85b87-52bzg\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.228217 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.286917 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jgtsh" event={"ID":"9d518023-a040-4289-b78b-163624c30b2a","Type":"ContainerStarted","Data":"25db876d869f6d0b28c7b4cd257b107c469d518aa40cf1436562b36f25d532b8"} Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.295709 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.355700 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" event={"ID":"7bba374b-c278-4218-917b-906554fffacd","Type":"ContainerDied","Data":"3c0fdb040cc6fcd32e64fc32914669eea738f194ca0a62220ee9256b40279c59"} Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.356639 4719 scope.go:117] "RemoveContainer" containerID="d96f4a08c4fc8b11192c93a930e3bfd07954f6026d9c01b8eb64232b220f0eb9" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.356838 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-xcjn4" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.355850 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-sb\") pod \"7bba374b-c278-4218-917b-906554fffacd\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.357432 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-svc\") pod \"7bba374b-c278-4218-917b-906554fffacd\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.357515 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-config\") pod \"7bba374b-c278-4218-917b-906554fffacd\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.357557 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffh4r\" (UniqueName: \"kubernetes.io/projected/7bba374b-c278-4218-917b-906554fffacd-kube-api-access-ffh4r\") pod \"7bba374b-c278-4218-917b-906554fffacd\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.357596 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-swift-storage-0\") pod \"7bba374b-c278-4218-917b-906554fffacd\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.357647 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-nb\") pod \"7bba374b-c278-4218-917b-906554fffacd\" (UID: \"7bba374b-c278-4218-917b-906554fffacd\") " Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.373098 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" event={"ID":"8370e556-0bec-4693-8440-644163299e5a","Type":"ContainerStarted","Data":"9135a3e4018c4807736653d80395cf5b78beaedf0abcc18e99a243b8ecebb40c"} Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.408237 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bba374b-c278-4218-917b-906554fffacd-kube-api-access-ffh4r" (OuterVolumeSpecName: "kube-api-access-ffh4r") pod "7bba374b-c278-4218-917b-906554fffacd" (UID: "7bba374b-c278-4218-917b-906554fffacd"). InnerVolumeSpecName "kube-api-access-ffh4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.433175 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.461284 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffh4r\" (UniqueName: \"kubernetes.io/projected/7bba374b-c278-4218-917b-906554fffacd-kube-api-access-ffh4r\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.527740 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7bba374b-c278-4218-917b-906554fffacd" (UID: "7bba374b-c278-4218-917b-906554fffacd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.528355 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-config" (OuterVolumeSpecName: "config") pod "7bba374b-c278-4218-917b-906554fffacd" (UID: "7bba374b-c278-4218-917b-906554fffacd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.585254 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.585323 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.585843 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7bba374b-c278-4218-917b-906554fffacd" (UID: "7bba374b-c278-4218-917b-906554fffacd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.643728 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5dc4d6cbb7-scwpt"] Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.663654 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9vwxh"] Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.681012 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.688023 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7bba374b-c278-4218-917b-906554fffacd" (UID: "7bba374b-c278-4218-917b-906554fffacd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.692664 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.692686 4719 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.705177 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7bba374b-c278-4218-917b-906554fffacd" (UID: "7bba374b-c278-4218-917b-906554fffacd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.751219 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gprbv"] Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.759269 4719 scope.go:117] "RemoveContainer" containerID="b1f1eaba2fa3f346a2dc54a03925118f0d23158e15d94315413446b28826ec8b" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.793966 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bba374b-c278-4218-917b-906554fffacd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:39 crc kubenswrapper[4719]: I1215 12:37:39.934047 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f46b45545-rn64s"] Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.004770 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-8qr2l"] Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.017924 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-xcjn4"] Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.051613 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-xcjn4"] Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.205663 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nd852"] Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.297565 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.398161 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edd905bf-6348-446d-abe8-fb632760bc1f","Type":"ContainerStarted","Data":"67f2f74f3a0bc116fef224d4f19e8313a697d86c2fc7d28016df2f048abdd183"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.398529 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-52bzg"] Dec 15 12:37:40 crc kubenswrapper[4719]: W1215 12:37:40.434408 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46738d80_d21e_43c5_8871_dba878a5986b.slice/crio-3253d1f65a34af251731c5573ca79f11ed3c5d016dee15dc886768ec72c84c78 WatchSource:0}: Error finding container 3253d1f65a34af251731c5573ca79f11ed3c5d016dee15dc886768ec72c84c78: Status 404 returned error can't find the container with id 3253d1f65a34af251731c5573ca79f11ed3c5d016dee15dc886768ec72c84c78 Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.434883 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nd852" event={"ID":"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c","Type":"ContainerStarted","Data":"b79d28b6888231612f46ec33c6ef828cd2656d4a0c24922933c5ac29e440d166"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.454894 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jgtsh" event={"ID":"9d518023-a040-4289-b78b-163624c30b2a","Type":"ContainerStarted","Data":"524d274eac51ccdfaa57b4292a48f01012b56b79adfbb9cfd91ef9a3471ee4eb"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.473709 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jgtsh" podStartSLOduration=3.473691733 podStartE2EDuration="3.473691733s" podCreationTimestamp="2025-12-15 12:37:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:40.472236208 +0000 UTC m=+1221.414529228" watchObservedRunningTime="2025-12-15 12:37:40.473691733 +0000 UTC m=+1221.415984763" Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.478983 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gprbv" event={"ID":"ef9ed6e9-3050-4335-8ffc-97f889eb63e6","Type":"ContainerStarted","Data":"d4ab5662a225660f3af235a30327f57713cd8d78e9122e59a3eba3233b4ddf73"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.504101 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8qr2l" event={"ID":"58fea5e6-9ff9-4a44-be8a-311735af3e4b","Type":"ContainerStarted","Data":"698165fdb097b663725a4116e1b32096409d5f48c974b38153507360e8bd7930"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.517120 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f46b45545-rn64s" event={"ID":"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4","Type":"ContainerStarted","Data":"0ae2c78d22e6fa1faf94da7f965488dc33332029ac10a20f1f7f0a1ee39a97a9"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.545997 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ece6b2ac-3524-402c-a20b-bfc65dc046b4","Type":"ContainerStarted","Data":"08cbd7725b9feedb5d525c16f80cc3d2f398456c903eb47d5d53c11113573719"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.551232 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9vwxh" event={"ID":"3243ab84-0a4c-4c74-83de-693c1c96e89e","Type":"ContainerStarted","Data":"f9a10627eba3e330e4bc0909a38d7d0e9a3da28f30c8c1e92e26e0c385194bd6"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.551275 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9vwxh" event={"ID":"3243ab84-0a4c-4c74-83de-693c1c96e89e","Type":"ContainerStarted","Data":"0ef71242f60fe5967eab9ab9f67dffd198606eaa844feb54153501caa6272b83"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.573204 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5dc4d6cbb7-scwpt" event={"ID":"27a897af-5729-4d7a-94a6-8318aa88eff5","Type":"ContainerStarted","Data":"9eb109d8aa3ce8963336a2282087fadd7ff10a6841a607234cfb123bd3481113"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.577458 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.632940 4719 generic.go:334] "Generic (PLEG): container finished" podID="8370e556-0bec-4693-8440-644163299e5a" containerID="ce6cd53c07f97bfba69de721570ed129454ad0fad6ce3d42f76a8e441009d086" exitCode=0 Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.633017 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" event={"ID":"8370e556-0bec-4693-8440-644163299e5a","Type":"ContainerDied","Data":"ce6cd53c07f97bfba69de721570ed129454ad0fad6ce3d42f76a8e441009d086"} Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.640095 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-9vwxh" podStartSLOduration=3.640074458 podStartE2EDuration="3.640074458s" podCreationTimestamp="2025-12-15 12:37:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:40.570147922 +0000 UTC m=+1221.512440952" watchObservedRunningTime="2025-12-15 12:37:40.640074458 +0000 UTC m=+1221.582367488" Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.969965 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5dc4d6cbb7-scwpt"] Dec 15 12:37:40 crc kubenswrapper[4719]: I1215 12:37:40.993956 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.021219 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-677f6b8777-tsxng"] Dec 15 12:37:41 crc kubenswrapper[4719]: E1215 12:37:41.021702 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bba374b-c278-4218-917b-906554fffacd" containerName="dnsmasq-dns" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.021729 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bba374b-c278-4218-917b-906554fffacd" containerName="dnsmasq-dns" Dec 15 12:37:41 crc kubenswrapper[4719]: E1215 12:37:41.021743 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bba374b-c278-4218-917b-906554fffacd" containerName="init" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.021750 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bba374b-c278-4218-917b-906554fffacd" containerName="init" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.021964 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bba374b-c278-4218-917b-906554fffacd" containerName="dnsmasq-dns" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.023070 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.087279 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-677f6b8777-tsxng"] Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.131140 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vpbs\" (UniqueName: \"kubernetes.io/projected/16a1675a-ef86-4636-bb06-dfd592019a53-kube-api-access-9vpbs\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.131201 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/16a1675a-ef86-4636-bb06-dfd592019a53-horizon-secret-key\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.131240 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a1675a-ef86-4636-bb06-dfd592019a53-logs\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.131264 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-config-data\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.131328 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-scripts\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.169000 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.233234 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-scripts\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.233328 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vpbs\" (UniqueName: \"kubernetes.io/projected/16a1675a-ef86-4636-bb06-dfd592019a53-kube-api-access-9vpbs\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.233381 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/16a1675a-ef86-4636-bb06-dfd592019a53-horizon-secret-key\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.233428 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a1675a-ef86-4636-bb06-dfd592019a53-logs\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.233455 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-config-data\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.235088 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.236144 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-scripts\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.236174 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a1675a-ef86-4636-bb06-dfd592019a53-logs\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.236679 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-config-data\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.258447 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/16a1675a-ef86-4636-bb06-dfd592019a53-horizon-secret-key\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.294432 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vpbs\" (UniqueName: \"kubernetes.io/projected/16a1675a-ef86-4636-bb06-dfd592019a53-kube-api-access-9vpbs\") pod \"horizon-677f6b8777-tsxng\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.309787 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.391984 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.438522 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-sb\") pod \"8370e556-0bec-4693-8440-644163299e5a\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.438660 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-swift-storage-0\") pod \"8370e556-0bec-4693-8440-644163299e5a\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.438707 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-svc\") pod \"8370e556-0bec-4693-8440-644163299e5a\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.438772 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-config\") pod \"8370e556-0bec-4693-8440-644163299e5a\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.438807 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwhdl\" (UniqueName: \"kubernetes.io/projected/8370e556-0bec-4693-8440-644163299e5a-kube-api-access-gwhdl\") pod \"8370e556-0bec-4693-8440-644163299e5a\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.438909 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-nb\") pod \"8370e556-0bec-4693-8440-644163299e5a\" (UID: \"8370e556-0bec-4693-8440-644163299e5a\") " Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.454204 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8370e556-0bec-4693-8440-644163299e5a-kube-api-access-gwhdl" (OuterVolumeSpecName: "kube-api-access-gwhdl") pod "8370e556-0bec-4693-8440-644163299e5a" (UID: "8370e556-0bec-4693-8440-644163299e5a"). InnerVolumeSpecName "kube-api-access-gwhdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.475354 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8370e556-0bec-4693-8440-644163299e5a" (UID: "8370e556-0bec-4693-8440-644163299e5a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.498590 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8370e556-0bec-4693-8440-644163299e5a" (UID: "8370e556-0bec-4693-8440-644163299e5a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.500283 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8370e556-0bec-4693-8440-644163299e5a" (UID: "8370e556-0bec-4693-8440-644163299e5a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.530156 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8370e556-0bec-4693-8440-644163299e5a" (UID: "8370e556-0bec-4693-8440-644163299e5a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.541290 4719 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.541325 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.541335 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwhdl\" (UniqueName: \"kubernetes.io/projected/8370e556-0bec-4693-8440-644163299e5a-kube-api-access-gwhdl\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.541344 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.541352 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.570723 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-config" (OuterVolumeSpecName: "config") pod "8370e556-0bec-4693-8440-644163299e5a" (UID: "8370e556-0bec-4693-8440-644163299e5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.619812 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bba374b-c278-4218-917b-906554fffacd" path="/var/lib/kubelet/pods/7bba374b-c278-4218-917b-906554fffacd/volumes" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.643341 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8370e556-0bec-4693-8440-644163299e5a-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.649673 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" event={"ID":"8370e556-0bec-4693-8440-644163299e5a","Type":"ContainerDied","Data":"9135a3e4018c4807736653d80395cf5b78beaedf0abcc18e99a243b8ecebb40c"} Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.649722 4719 scope.go:117] "RemoveContainer" containerID="ce6cd53c07f97bfba69de721570ed129454ad0fad6ce3d42f76a8e441009d086" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.649969 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl" Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.667207 4719 generic.go:334] "Generic (PLEG): container finished" podID="46738d80-d21e-43c5-8871-dba878a5986b" containerID="4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8" exitCode=0 Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.667344 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" event={"ID":"46738d80-d21e-43c5-8871-dba878a5986b","Type":"ContainerDied","Data":"4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8"} Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.667391 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" event={"ID":"46738d80-d21e-43c5-8871-dba878a5986b","Type":"ContainerStarted","Data":"3253d1f65a34af251731c5573ca79f11ed3c5d016dee15dc886768ec72c84c78"} Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.671051 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5e309330-aa51-4178-a5cb-2f1ae1255255","Type":"ContainerStarted","Data":"5a99584cac6b4f327f93481f57a4abf4791a2b9b2b66e21bfcfc56bc3edd65d5"} Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.702150 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl"] Dec 15 12:37:41 crc kubenswrapper[4719]: I1215 12:37:41.708932 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8r8vl"] Dec 15 12:37:42 crc kubenswrapper[4719]: I1215 12:37:42.196878 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-677f6b8777-tsxng"] Dec 15 12:37:42 crc kubenswrapper[4719]: W1215 12:37:42.221291 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16a1675a_ef86_4636_bb06_dfd592019a53.slice/crio-bf5685a8b7ce2f086b8b97d7a354e0c15af7570b634054a3cf21951c20f885d2 WatchSource:0}: Error finding container bf5685a8b7ce2f086b8b97d7a354e0c15af7570b634054a3cf21951c20f885d2: Status 404 returned error can't find the container with id bf5685a8b7ce2f086b8b97d7a354e0c15af7570b634054a3cf21951c20f885d2 Dec 15 12:37:42 crc kubenswrapper[4719]: I1215 12:37:42.729120 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" event={"ID":"46738d80-d21e-43c5-8871-dba878a5986b","Type":"ContainerStarted","Data":"3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d"} Dec 15 12:37:42 crc kubenswrapper[4719]: I1215 12:37:42.734089 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:42 crc kubenswrapper[4719]: I1215 12:37:42.743523 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-677f6b8777-tsxng" event={"ID":"16a1675a-ef86-4636-bb06-dfd592019a53","Type":"ContainerStarted","Data":"bf5685a8b7ce2f086b8b97d7a354e0c15af7570b634054a3cf21951c20f885d2"} Dec 15 12:37:42 crc kubenswrapper[4719]: I1215 12:37:42.788623 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" podStartSLOduration=4.78860434 podStartE2EDuration="4.78860434s" podCreationTimestamp="2025-12-15 12:37:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:42.785559935 +0000 UTC m=+1223.727852965" watchObservedRunningTime="2025-12-15 12:37:42.78860434 +0000 UTC m=+1223.730897370" Dec 15 12:37:42 crc kubenswrapper[4719]: I1215 12:37:42.814435 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ece6b2ac-3524-402c-a20b-bfc65dc046b4","Type":"ContainerStarted","Data":"c5bce032c1f3067459d042d78d4b8b6aeaa3d12373f285a6312aff0d01f12fc3"} Dec 15 12:37:42 crc kubenswrapper[4719]: I1215 12:37:42.823947 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5e309330-aa51-4178-a5cb-2f1ae1255255","Type":"ContainerStarted","Data":"93daf2ea66ab371ecf3c3b31844e8f1700c620068a05444febad6f71febba7da"} Dec 15 12:37:43 crc kubenswrapper[4719]: I1215 12:37:43.490387 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8370e556-0bec-4693-8440-644163299e5a" path="/var/lib/kubelet/pods/8370e556-0bec-4693-8440-644163299e5a/volumes" Dec 15 12:37:43 crc kubenswrapper[4719]: I1215 12:37:43.840899 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ece6b2ac-3524-402c-a20b-bfc65dc046b4","Type":"ContainerStarted","Data":"f6098831bee47021d20b2d4a4165d88771f204deae2ba4df2d2be39458ec51a6"} Dec 15 12:37:43 crc kubenswrapper[4719]: I1215 12:37:43.840944 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" containerName="glance-log" containerID="cri-o://c5bce032c1f3067459d042d78d4b8b6aeaa3d12373f285a6312aff0d01f12fc3" gracePeriod=30 Dec 15 12:37:43 crc kubenswrapper[4719]: I1215 12:37:43.841341 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" containerName="glance-httpd" containerID="cri-o://f6098831bee47021d20b2d4a4165d88771f204deae2ba4df2d2be39458ec51a6" gracePeriod=30 Dec 15 12:37:43 crc kubenswrapper[4719]: I1215 12:37:43.864074 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.864055224 podStartE2EDuration="5.864055224s" podCreationTimestamp="2025-12-15 12:37:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:43.859454081 +0000 UTC m=+1224.801747111" watchObservedRunningTime="2025-12-15 12:37:43.864055224 +0000 UTC m=+1224.806348254" Dec 15 12:37:44 crc kubenswrapper[4719]: I1215 12:37:44.859824 4719 generic.go:334] "Generic (PLEG): container finished" podID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" containerID="c5bce032c1f3067459d042d78d4b8b6aeaa3d12373f285a6312aff0d01f12fc3" exitCode=143 Dec 15 12:37:44 crc kubenswrapper[4719]: I1215 12:37:44.860802 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ece6b2ac-3524-402c-a20b-bfc65dc046b4","Type":"ContainerDied","Data":"c5bce032c1f3067459d042d78d4b8b6aeaa3d12373f285a6312aff0d01f12fc3"} Dec 15 12:37:45 crc kubenswrapper[4719]: I1215 12:37:45.897361 4719 generic.go:334] "Generic (PLEG): container finished" podID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" containerID="f6098831bee47021d20b2d4a4165d88771f204deae2ba4df2d2be39458ec51a6" exitCode=0 Dec 15 12:37:45 crc kubenswrapper[4719]: I1215 12:37:45.897531 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ece6b2ac-3524-402c-a20b-bfc65dc046b4","Type":"ContainerDied","Data":"f6098831bee47021d20b2d4a4165d88771f204deae2ba4df2d2be39458ec51a6"} Dec 15 12:37:45 crc kubenswrapper[4719]: I1215 12:37:45.899968 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5e309330-aa51-4178-a5cb-2f1ae1255255","Type":"ContainerStarted","Data":"103948537bba25ac67d5771374cc139f4b1d0eceeb003de13c0219aae21ecb6c"} Dec 15 12:37:45 crc kubenswrapper[4719]: I1215 12:37:45.900089 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5e309330-aa51-4178-a5cb-2f1ae1255255" containerName="glance-log" containerID="cri-o://93daf2ea66ab371ecf3c3b31844e8f1700c620068a05444febad6f71febba7da" gracePeriod=30 Dec 15 12:37:45 crc kubenswrapper[4719]: I1215 12:37:45.900546 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5e309330-aa51-4178-a5cb-2f1ae1255255" containerName="glance-httpd" containerID="cri-o://103948537bba25ac67d5771374cc139f4b1d0eceeb003de13c0219aae21ecb6c" gracePeriod=30 Dec 15 12:37:45 crc kubenswrapper[4719]: I1215 12:37:45.940381 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.94036366 podStartE2EDuration="7.94036366s" podCreationTimestamp="2025-12-15 12:37:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:37:45.933394694 +0000 UTC m=+1226.875687724" watchObservedRunningTime="2025-12-15 12:37:45.94036366 +0000 UTC m=+1226.882656690" Dec 15 12:37:46 crc kubenswrapper[4719]: E1215 12:37:46.226806 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e309330_aa51_4178_a5cb_2f1ae1255255.slice/crio-103948537bba25ac67d5771374cc139f4b1d0eceeb003de13c0219aae21ecb6c.scope\": RecentStats: unable to find data in memory cache]" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.279152 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.478639 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-combined-ca-bundle\") pod \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.478975 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-public-tls-certs\") pod \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.479666 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-httpd-run\") pod \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.480327 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m64z\" (UniqueName: \"kubernetes.io/projected/ece6b2ac-3524-402c-a20b-bfc65dc046b4-kube-api-access-9m64z\") pod \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.480371 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-config-data\") pod \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.480628 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-logs\") pod \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.480666 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-scripts\") pod \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.480722 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\" (UID: \"ece6b2ac-3524-402c-a20b-bfc65dc046b4\") " Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.482171 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-logs" (OuterVolumeSpecName: "logs") pod "ece6b2ac-3524-402c-a20b-bfc65dc046b4" (UID: "ece6b2ac-3524-402c-a20b-bfc65dc046b4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.482711 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ece6b2ac-3524-402c-a20b-bfc65dc046b4" (UID: "ece6b2ac-3524-402c-a20b-bfc65dc046b4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.487201 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ece6b2ac-3524-402c-a20b-bfc65dc046b4-kube-api-access-9m64z" (OuterVolumeSpecName: "kube-api-access-9m64z") pod "ece6b2ac-3524-402c-a20b-bfc65dc046b4" (UID: "ece6b2ac-3524-402c-a20b-bfc65dc046b4"). InnerVolumeSpecName "kube-api-access-9m64z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.490707 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-scripts" (OuterVolumeSpecName: "scripts") pod "ece6b2ac-3524-402c-a20b-bfc65dc046b4" (UID: "ece6b2ac-3524-402c-a20b-bfc65dc046b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.509751 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "ece6b2ac-3524-402c-a20b-bfc65dc046b4" (UID: "ece6b2ac-3524-402c-a20b-bfc65dc046b4"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.533970 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ece6b2ac-3524-402c-a20b-bfc65dc046b4" (UID: "ece6b2ac-3524-402c-a20b-bfc65dc046b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.537600 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-config-data" (OuterVolumeSpecName: "config-data") pod "ece6b2ac-3524-402c-a20b-bfc65dc046b4" (UID: "ece6b2ac-3524-402c-a20b-bfc65dc046b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.552200 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ece6b2ac-3524-402c-a20b-bfc65dc046b4" (UID: "ece6b2ac-3524-402c-a20b-bfc65dc046b4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.582319 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.582349 4719 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.582359 4719 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.582368 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m64z\" (UniqueName: \"kubernetes.io/projected/ece6b2ac-3524-402c-a20b-bfc65dc046b4-kube-api-access-9m64z\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.582379 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.582386 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ece6b2ac-3524-402c-a20b-bfc65dc046b4-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.582394 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ece6b2ac-3524-402c-a20b-bfc65dc046b4-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.582425 4719 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.600031 4719 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.683430 4719 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.910863 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f46b45545-rn64s"] Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.960970 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ece6b2ac-3524-402c-a20b-bfc65dc046b4","Type":"ContainerDied","Data":"08cbd7725b9feedb5d525c16f80cc3d2f398456c903eb47d5d53c11113573719"} Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.961022 4719 scope.go:117] "RemoveContainer" containerID="f6098831bee47021d20b2d4a4165d88771f204deae2ba4df2d2be39458ec51a6" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.961178 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.970389 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c6bcf58b6-vx99l"] Dec 15 12:37:46 crc kubenswrapper[4719]: E1215 12:37:46.970761 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8370e556-0bec-4693-8440-644163299e5a" containerName="init" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.970780 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="8370e556-0bec-4693-8440-644163299e5a" containerName="init" Dec 15 12:37:46 crc kubenswrapper[4719]: E1215 12:37:46.970797 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" containerName="glance-httpd" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.970804 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" containerName="glance-httpd" Dec 15 12:37:46 crc kubenswrapper[4719]: E1215 12:37:46.970814 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" containerName="glance-log" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.970820 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" containerName="glance-log" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.970981 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" containerName="glance-log" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.971004 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="8370e556-0bec-4693-8440-644163299e5a" containerName="init" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.971019 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" containerName="glance-httpd" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.971828 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.976889 4719 generic.go:334] "Generic (PLEG): container finished" podID="5e309330-aa51-4178-a5cb-2f1ae1255255" containerID="103948537bba25ac67d5771374cc139f4b1d0eceeb003de13c0219aae21ecb6c" exitCode=0 Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.977078 4719 generic.go:334] "Generic (PLEG): container finished" podID="5e309330-aa51-4178-a5cb-2f1ae1255255" containerID="93daf2ea66ab371ecf3c3b31844e8f1700c620068a05444febad6f71febba7da" exitCode=143 Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.977153 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5e309330-aa51-4178-a5cb-2f1ae1255255","Type":"ContainerDied","Data":"103948537bba25ac67d5771374cc139f4b1d0eceeb003de13c0219aae21ecb6c"} Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.977229 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5e309330-aa51-4178-a5cb-2f1ae1255255","Type":"ContainerDied","Data":"93daf2ea66ab371ecf3c3b31844e8f1700c620068a05444febad6f71febba7da"} Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.978001 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.987156 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-scripts\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.987202 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-config-data\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.987250 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-tls-certs\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.987270 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-secret-key\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.987293 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8b2b\" (UniqueName: \"kubernetes.io/projected/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-kube-api-access-b8b2b\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.987315 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-combined-ca-bundle\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:46 crc kubenswrapper[4719]: I1215 12:37:46.987361 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-logs\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.003294 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c6bcf58b6-vx99l"] Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.080845 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-677f6b8777-tsxng"] Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.088663 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-config-data\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.088761 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-tls-certs\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.088787 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-secret-key\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.088822 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8b2b\" (UniqueName: \"kubernetes.io/projected/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-kube-api-access-b8b2b\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.088873 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-combined-ca-bundle\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.088930 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-logs\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.088983 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-scripts\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.089782 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-scripts\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.090468 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-config-data\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.090735 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-logs\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.123099 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-combined-ca-bundle\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.123694 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-secret-key\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.128359 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-tls-certs\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.155915 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.164509 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8b2b\" (UniqueName: \"kubernetes.io/projected/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-kube-api-access-b8b2b\") pod \"horizon-7c6bcf58b6-vx99l\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.183224 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6bc5fb48-jqf6b"] Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.184961 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.202822 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.222439 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bc5fb48-jqf6b"] Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.233748 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.249346 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.254437 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.265146 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.293481 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.303917 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-horizon-secret-key\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.304004 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-config-data\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.304057 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm9l7\" (UniqueName: \"kubernetes.io/projected/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-kube-api-access-lm9l7\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.304089 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-scripts\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.304277 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-combined-ca-bundle\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.304305 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-horizon-tls-certs\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.304415 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-logs\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.388387 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406440 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-combined-ca-bundle\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406494 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-horizon-tls-certs\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406525 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406559 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406611 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-logs\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406633 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406687 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-horizon-secret-key\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406709 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-logs\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406745 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-config-data\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406774 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm9l7\" (UniqueName: \"kubernetes.io/projected/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-kube-api-access-lm9l7\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406799 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406823 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-scripts\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406887 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406912 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbj2l\" (UniqueName: \"kubernetes.io/projected/63366e84-551e-4eb0-a007-7156b6ab1d6b-kube-api-access-tbj2l\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.406930 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.408638 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-logs\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.409052 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-config-data\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.411338 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-scripts\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.420651 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-horizon-secret-key\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.436055 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-combined-ca-bundle\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.442725 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm9l7\" (UniqueName: \"kubernetes.io/projected/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-kube-api-access-lm9l7\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.443627 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926-horizon-tls-certs\") pod \"horizon-6bc5fb48-jqf6b\" (UID: \"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926\") " pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.509407 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.509451 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbj2l\" (UniqueName: \"kubernetes.io/projected/63366e84-551e-4eb0-a007-7156b6ab1d6b-kube-api-access-tbj2l\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.509470 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.509553 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.509574 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.509636 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.509694 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-logs\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.509730 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.510313 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.510684 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.515118 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.515675 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.510702 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-logs\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.526966 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.527836 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.531661 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbj2l\" (UniqueName: \"kubernetes.io/projected/63366e84-551e-4eb0-a007-7156b6ab1d6b-kube-api-access-tbj2l\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.539973 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " pod="openstack/glance-default-external-api-0" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.541990 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ece6b2ac-3524-402c-a20b-bfc65dc046b4" path="/var/lib/kubelet/pods/ece6b2ac-3524-402c-a20b-bfc65dc046b4/volumes" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.574125 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:37:47 crc kubenswrapper[4719]: I1215 12:37:47.597539 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 15 12:37:49 crc kubenswrapper[4719]: I1215 12:37:49.298372 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:37:49 crc kubenswrapper[4719]: I1215 12:37:49.365878 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-m9j55"] Dec 15 12:37:49 crc kubenswrapper[4719]: I1215 12:37:49.366162 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="dnsmasq-dns" containerID="cri-o://0e96ac6e3ea981f737bec1c9446e922db723f7ab9152cce21ff8a3f7e3d80063" gracePeriod=10 Dec 15 12:37:50 crc kubenswrapper[4719]: I1215 12:37:50.024481 4719 generic.go:334] "Generic (PLEG): container finished" podID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerID="0e96ac6e3ea981f737bec1c9446e922db723f7ab9152cce21ff8a3f7e3d80063" exitCode=0 Dec 15 12:37:50 crc kubenswrapper[4719]: I1215 12:37:50.024558 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" event={"ID":"9bb0baa8-ce19-4923-9e24-66933db4c618","Type":"ContainerDied","Data":"0e96ac6e3ea981f737bec1c9446e922db723f7ab9152cce21ff8a3f7e3d80063"} Dec 15 12:37:51 crc kubenswrapper[4719]: I1215 12:37:51.983640 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.055602 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5e309330-aa51-4178-a5cb-2f1ae1255255","Type":"ContainerDied","Data":"5a99584cac6b4f327f93481f57a4abf4791a2b9b2b66e21bfcfc56bc3edd65d5"} Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.055810 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.099292 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pmzr\" (UniqueName: \"kubernetes.io/projected/5e309330-aa51-4178-a5cb-2f1ae1255255-kube-api-access-9pmzr\") pod \"5e309330-aa51-4178-a5cb-2f1ae1255255\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.099358 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-config-data\") pod \"5e309330-aa51-4178-a5cb-2f1ae1255255\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.099393 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-internal-tls-certs\") pod \"5e309330-aa51-4178-a5cb-2f1ae1255255\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.099455 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-httpd-run\") pod \"5e309330-aa51-4178-a5cb-2f1ae1255255\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.099571 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-combined-ca-bundle\") pod \"5e309330-aa51-4178-a5cb-2f1ae1255255\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.099612 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-scripts\") pod \"5e309330-aa51-4178-a5cb-2f1ae1255255\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.099666 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-logs\") pod \"5e309330-aa51-4178-a5cb-2f1ae1255255\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.099705 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"5e309330-aa51-4178-a5cb-2f1ae1255255\" (UID: \"5e309330-aa51-4178-a5cb-2f1ae1255255\") " Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.100275 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-logs" (OuterVolumeSpecName: "logs") pod "5e309330-aa51-4178-a5cb-2f1ae1255255" (UID: "5e309330-aa51-4178-a5cb-2f1ae1255255"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.100292 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5e309330-aa51-4178-a5cb-2f1ae1255255" (UID: "5e309330-aa51-4178-a5cb-2f1ae1255255"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.105792 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "5e309330-aa51-4178-a5cb-2f1ae1255255" (UID: "5e309330-aa51-4178-a5cb-2f1ae1255255"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.106313 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e309330-aa51-4178-a5cb-2f1ae1255255-kube-api-access-9pmzr" (OuterVolumeSpecName: "kube-api-access-9pmzr") pod "5e309330-aa51-4178-a5cb-2f1ae1255255" (UID: "5e309330-aa51-4178-a5cb-2f1ae1255255"). InnerVolumeSpecName "kube-api-access-9pmzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.110405 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-scripts" (OuterVolumeSpecName: "scripts") pod "5e309330-aa51-4178-a5cb-2f1ae1255255" (UID: "5e309330-aa51-4178-a5cb-2f1ae1255255"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.172033 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-config-data" (OuterVolumeSpecName: "config-data") pod "5e309330-aa51-4178-a5cb-2f1ae1255255" (UID: "5e309330-aa51-4178-a5cb-2f1ae1255255"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.178053 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5e309330-aa51-4178-a5cb-2f1ae1255255" (UID: "5e309330-aa51-4178-a5cb-2f1ae1255255"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.179004 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e309330-aa51-4178-a5cb-2f1ae1255255" (UID: "5e309330-aa51-4178-a5cb-2f1ae1255255"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.202086 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.202403 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.202417 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.202452 4719 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.202463 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pmzr\" (UniqueName: \"kubernetes.io/projected/5e309330-aa51-4178-a5cb-2f1ae1255255-kube-api-access-9pmzr\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.202475 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.202516 4719 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e309330-aa51-4178-a5cb-2f1ae1255255-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.202526 4719 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5e309330-aa51-4178-a5cb-2f1ae1255255-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.233912 4719 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.304268 4719 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.399151 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.418708 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.433914 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:37:52 crc kubenswrapper[4719]: E1215 12:37:52.434373 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e309330-aa51-4178-a5cb-2f1ae1255255" containerName="glance-log" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.434395 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e309330-aa51-4178-a5cb-2f1ae1255255" containerName="glance-log" Dec 15 12:37:52 crc kubenswrapper[4719]: E1215 12:37:52.434431 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e309330-aa51-4178-a5cb-2f1ae1255255" containerName="glance-httpd" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.434440 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e309330-aa51-4178-a5cb-2f1ae1255255" containerName="glance-httpd" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.434680 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e309330-aa51-4178-a5cb-2f1ae1255255" containerName="glance-log" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.434718 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e309330-aa51-4178-a5cb-2f1ae1255255" containerName="glance-httpd" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.435810 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.440572 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.440772 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.448048 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.510773 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-logs\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.510830 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.510880 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.510918 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-config-data\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.510937 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.510972 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-scripts\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.510995 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.511042 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dghkz\" (UniqueName: \"kubernetes.io/projected/51211ffc-0363-4e48-9972-15b8e997d4de-kube-api-access-dghkz\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.614016 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-config-data\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.614079 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.614137 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-scripts\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.614168 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.614241 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dghkz\" (UniqueName: \"kubernetes.io/projected/51211ffc-0363-4e48-9972-15b8e997d4de-kube-api-access-dghkz\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.614288 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-logs\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.614326 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.614367 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.617781 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.843305 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-logs\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.843726 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.846679 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.846868 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dghkz\" (UniqueName: \"kubernetes.io/projected/51211ffc-0363-4e48-9972-15b8e997d4de-kube-api-access-dghkz\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.847819 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-scripts\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.848765 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-config-data\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.848957 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:52 crc kubenswrapper[4719]: I1215 12:37:52.851412 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:37:53 crc kubenswrapper[4719]: I1215 12:37:53.144314 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 15 12:37:53 crc kubenswrapper[4719]: I1215 12:37:53.385771 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Dec 15 12:37:53 crc kubenswrapper[4719]: I1215 12:37:53.486579 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e309330-aa51-4178-a5cb-2f1ae1255255" path="/var/lib/kubelet/pods/5e309330-aa51-4178-a5cb-2f1ae1255255/volumes" Dec 15 12:37:55 crc kubenswrapper[4719]: I1215 12:37:55.076217 4719 generic.go:334] "Generic (PLEG): container finished" podID="9d518023-a040-4289-b78b-163624c30b2a" containerID="524d274eac51ccdfaa57b4292a48f01012b56b79adfbb9cfd91ef9a3471ee4eb" exitCode=0 Dec 15 12:37:55 crc kubenswrapper[4719]: I1215 12:37:55.076426 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jgtsh" event={"ID":"9d518023-a040-4289-b78b-163624c30b2a","Type":"ContainerDied","Data":"524d274eac51ccdfaa57b4292a48f01012b56b79adfbb9cfd91ef9a3471ee4eb"} Dec 15 12:37:58 crc kubenswrapper[4719]: I1215 12:37:58.385616 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Dec 15 12:37:59 crc kubenswrapper[4719]: E1215 12:37:59.353714 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 15 12:37:59 crc kubenswrapper[4719]: E1215 12:37:59.353973 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nbbh674h5b8h695h7h68bhb5h585h67fh55dh588h7dh585h57h548h55fh55h59dh9dh5d7h5d9h5cbh558h59h5c5h684h86h5cbh84hd4h4h598q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9vpbs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-677f6b8777-tsxng_openstack(16a1675a-ef86-4636-bb06-dfd592019a53): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:37:59 crc kubenswrapper[4719]: E1215 12:37:59.357742 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-677f6b8777-tsxng" podUID="16a1675a-ef86-4636-bb06-dfd592019a53" Dec 15 12:38:02 crc kubenswrapper[4719]: E1215 12:38:02.341586 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 15 12:38:02 crc kubenswrapper[4719]: E1215 12:38:02.344969 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8dh578hdfhb6h686h55ch54ch648hb8h585h547h78h56chb9h5d4h686h6fhcdh64dh84hc4h5ch95h8ch56dh54dh5bh5fh66ch58bhchd6q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rmmgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(edd905bf-6348-446d-abe8-fb632760bc1f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:38:06 crc kubenswrapper[4719]: E1215 12:38:06.665237 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 15 12:38:06 crc kubenswrapper[4719]: E1215 12:38:06.667114 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rlds7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-nd852_openstack(cba0d08d-bba6-4a68-bc44-c4cf1619ad4c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:38:06 crc kubenswrapper[4719]: E1215 12:38:06.669182 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-nd852" podUID="cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" Dec 15 12:38:06 crc kubenswrapper[4719]: E1215 12:38:06.677175 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 15 12:38:06 crc kubenswrapper[4719]: E1215 12:38:06.677328 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9h679hfdh55dh5b9h587h687h6bh9ch5c8h5fch5hb9h77h55h67bh68ch5fbh67ch57h668h678hd5h588h54bh5cfhcch664h555h4h67h6cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7fbpq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5dc4d6cbb7-scwpt_openstack(27a897af-5729-4d7a-94a6-8318aa88eff5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:38:06 crc kubenswrapper[4719]: E1215 12:38:06.681798 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5dc4d6cbb7-scwpt" podUID="27a897af-5729-4d7a-94a6-8318aa88eff5" Dec 15 12:38:06 crc kubenswrapper[4719]: E1215 12:38:06.703020 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 15 12:38:06 crc kubenswrapper[4719]: E1215 12:38:06.703176 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh6dh585h675h697h5f6h84h8bh565h5bfh557hdbh74h594h664h69hd9h577hc9hbh645h85h5d7h97h5c8h596h65bh54fh5cdh584h6hf4q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hv4vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5f46b45545-rn64s_openstack(7350a4c2-fbaf-44d2-afd4-8001bc71b2a4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:38:06 crc kubenswrapper[4719]: E1215 12:38:06.707117 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5f46b45545-rn64s" podUID="7350a4c2-fbaf-44d2-afd4-8001bc71b2a4" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.730963 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.780897 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-scripts\") pod \"9d518023-a040-4289-b78b-163624c30b2a\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.780951 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-combined-ca-bundle\") pod \"9d518023-a040-4289-b78b-163624c30b2a\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.781032 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzz67\" (UniqueName: \"kubernetes.io/projected/9d518023-a040-4289-b78b-163624c30b2a-kube-api-access-zzz67\") pod \"9d518023-a040-4289-b78b-163624c30b2a\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.781136 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-credential-keys\") pod \"9d518023-a040-4289-b78b-163624c30b2a\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.781155 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-fernet-keys\") pod \"9d518023-a040-4289-b78b-163624c30b2a\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.781203 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-config-data\") pod \"9d518023-a040-4289-b78b-163624c30b2a\" (UID: \"9d518023-a040-4289-b78b-163624c30b2a\") " Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.792280 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-scripts" (OuterVolumeSpecName: "scripts") pod "9d518023-a040-4289-b78b-163624c30b2a" (UID: "9d518023-a040-4289-b78b-163624c30b2a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.793747 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9d518023-a040-4289-b78b-163624c30b2a" (UID: "9d518023-a040-4289-b78b-163624c30b2a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.794482 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9d518023-a040-4289-b78b-163624c30b2a" (UID: "9d518023-a040-4289-b78b-163624c30b2a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.794736 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d518023-a040-4289-b78b-163624c30b2a-kube-api-access-zzz67" (OuterVolumeSpecName: "kube-api-access-zzz67") pod "9d518023-a040-4289-b78b-163624c30b2a" (UID: "9d518023-a040-4289-b78b-163624c30b2a"). InnerVolumeSpecName "kube-api-access-zzz67". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.818148 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-config-data" (OuterVolumeSpecName: "config-data") pod "9d518023-a040-4289-b78b-163624c30b2a" (UID: "9d518023-a040-4289-b78b-163624c30b2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.826272 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d518023-a040-4289-b78b-163624c30b2a" (UID: "9d518023-a040-4289-b78b-163624c30b2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.884022 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzz67\" (UniqueName: \"kubernetes.io/projected/9d518023-a040-4289-b78b-163624c30b2a-kube-api-access-zzz67\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.884342 4719 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.884352 4719 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.884363 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.884372 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:06 crc kubenswrapper[4719]: I1215 12:38:06.884381 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d518023-a040-4289-b78b-163624c30b2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.183778 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jgtsh" event={"ID":"9d518023-a040-4289-b78b-163624c30b2a","Type":"ContainerDied","Data":"25db876d869f6d0b28c7b4cd257b107c469d518aa40cf1436562b36f25d532b8"} Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.183830 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25db876d869f6d0b28c7b4cd257b107c469d518aa40cf1436562b36f25d532b8" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.184962 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jgtsh" Dec 15 12:38:07 crc kubenswrapper[4719]: E1215 12:38:07.187908 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-nd852" podUID="cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.814297 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jgtsh"] Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.821075 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jgtsh"] Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.923978 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-88rcn"] Dec 15 12:38:07 crc kubenswrapper[4719]: E1215 12:38:07.924501 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d518023-a040-4289-b78b-163624c30b2a" containerName="keystone-bootstrap" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.924524 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d518023-a040-4289-b78b-163624c30b2a" containerName="keystone-bootstrap" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.924753 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d518023-a040-4289-b78b-163624c30b2a" containerName="keystone-bootstrap" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.925492 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.928384 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.928690 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.928887 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.929083 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kdx8k" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.929415 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 15 12:38:07 crc kubenswrapper[4719]: I1215 12:38:07.953549 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-88rcn"] Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.018988 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-config-data\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.019053 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pzzj\" (UniqueName: \"kubernetes.io/projected/978132fa-fca3-4665-b133-682da5d02f93-kube-api-access-4pzzj\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.019103 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-credential-keys\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.019175 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-scripts\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.019198 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-combined-ca-bundle\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.019221 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-fernet-keys\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.121291 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-credential-keys\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.121407 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-scripts\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.121434 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-combined-ca-bundle\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.121460 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-fernet-keys\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.121592 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-config-data\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.121634 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pzzj\" (UniqueName: \"kubernetes.io/projected/978132fa-fca3-4665-b133-682da5d02f93-kube-api-access-4pzzj\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.127804 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-scripts\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.131181 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-credential-keys\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.132473 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-fernet-keys\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.132522 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-config-data\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.139674 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pzzj\" (UniqueName: \"kubernetes.io/projected/978132fa-fca3-4665-b133-682da5d02f93-kube-api-access-4pzzj\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.144301 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-combined-ca-bundle\") pod \"keystone-bootstrap-88rcn\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.250225 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.386590 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 15 12:38:08 crc kubenswrapper[4719]: I1215 12:38:08.386836 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:38:09 crc kubenswrapper[4719]: I1215 12:38:09.486720 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d518023-a040-4289-b78b-163624c30b2a" path="/var/lib/kubelet/pods/9d518023-a040-4289-b78b-163624c30b2a/volumes" Dec 15 12:38:13 crc kubenswrapper[4719]: I1215 12:38:13.387809 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 15 12:38:18 crc kubenswrapper[4719]: I1215 12:38:18.389118 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.475087 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.561408 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/16a1675a-ef86-4636-bb06-dfd592019a53-horizon-secret-key\") pod \"16a1675a-ef86-4636-bb06-dfd592019a53\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.561721 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-scripts\") pod \"16a1675a-ef86-4636-bb06-dfd592019a53\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.562112 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-scripts" (OuterVolumeSpecName: "scripts") pod "16a1675a-ef86-4636-bb06-dfd592019a53" (UID: "16a1675a-ef86-4636-bb06-dfd592019a53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.562238 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-config-data\") pod \"16a1675a-ef86-4636-bb06-dfd592019a53\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.562343 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vpbs\" (UniqueName: \"kubernetes.io/projected/16a1675a-ef86-4636-bb06-dfd592019a53-kube-api-access-9vpbs\") pod \"16a1675a-ef86-4636-bb06-dfd592019a53\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.562395 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a1675a-ef86-4636-bb06-dfd592019a53-logs\") pod \"16a1675a-ef86-4636-bb06-dfd592019a53\" (UID: \"16a1675a-ef86-4636-bb06-dfd592019a53\") " Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.562815 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16a1675a-ef86-4636-bb06-dfd592019a53-logs" (OuterVolumeSpecName: "logs") pod "16a1675a-ef86-4636-bb06-dfd592019a53" (UID: "16a1675a-ef86-4636-bb06-dfd592019a53"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.562930 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.564284 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-config-data" (OuterVolumeSpecName: "config-data") pod "16a1675a-ef86-4636-bb06-dfd592019a53" (UID: "16a1675a-ef86-4636-bb06-dfd592019a53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.576319 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16a1675a-ef86-4636-bb06-dfd592019a53-kube-api-access-9vpbs" (OuterVolumeSpecName: "kube-api-access-9vpbs") pod "16a1675a-ef86-4636-bb06-dfd592019a53" (UID: "16a1675a-ef86-4636-bb06-dfd592019a53"). InnerVolumeSpecName "kube-api-access-9vpbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.576664 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16a1675a-ef86-4636-bb06-dfd592019a53-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "16a1675a-ef86-4636-bb06-dfd592019a53" (UID: "16a1675a-ef86-4636-bb06-dfd592019a53"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.664891 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16a1675a-ef86-4636-bb06-dfd592019a53-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.664941 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vpbs\" (UniqueName: \"kubernetes.io/projected/16a1675a-ef86-4636-bb06-dfd592019a53-kube-api-access-9vpbs\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.664952 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16a1675a-ef86-4636-bb06-dfd592019a53-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.664960 4719 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/16a1675a-ef86-4636-bb06-dfd592019a53-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:20 crc kubenswrapper[4719]: E1215 12:38:20.942710 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 15 12:38:20 crc kubenswrapper[4719]: E1215 12:38:20.942894 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fxzg5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-8qr2l_openstack(58fea5e6-9ff9-4a44-be8a-311735af3e4b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:38:20 crc kubenswrapper[4719]: E1215 12:38:20.944092 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-8qr2l" podUID="58fea5e6-9ff9-4a44-be8a-311735af3e4b" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.958253 4719 scope.go:117] "RemoveContainer" containerID="c5bce032c1f3067459d042d78d4b8b6aeaa3d12373f285a6312aff0d01f12fc3" Dec 15 12:38:20 crc kubenswrapper[4719]: I1215 12:38:20.993602 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.023789 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.038758 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071566 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-config\") pod \"9bb0baa8-ce19-4923-9e24-66933db4c618\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071606 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27a897af-5729-4d7a-94a6-8318aa88eff5-horizon-secret-key\") pod \"27a897af-5729-4d7a-94a6-8318aa88eff5\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071651 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-nb\") pod \"9bb0baa8-ce19-4923-9e24-66933db4c618\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071678 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-horizon-secret-key\") pod \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071706 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-logs\") pod \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071748 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fbpq\" (UniqueName: \"kubernetes.io/projected/27a897af-5729-4d7a-94a6-8318aa88eff5-kube-api-access-7fbpq\") pod \"27a897af-5729-4d7a-94a6-8318aa88eff5\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071774 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-config-data\") pod \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071805 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv4vq\" (UniqueName: \"kubernetes.io/projected/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-kube-api-access-hv4vq\") pod \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071829 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-config-data\") pod \"27a897af-5729-4d7a-94a6-8318aa88eff5\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071876 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a897af-5729-4d7a-94a6-8318aa88eff5-logs\") pod \"27a897af-5729-4d7a-94a6-8318aa88eff5\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.071899 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-scripts\") pod \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\" (UID: \"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.072032 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-dns-svc\") pod \"9bb0baa8-ce19-4923-9e24-66933db4c618\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.072079 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-scripts\") pod \"27a897af-5729-4d7a-94a6-8318aa88eff5\" (UID: \"27a897af-5729-4d7a-94a6-8318aa88eff5\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.072101 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx5bh\" (UniqueName: \"kubernetes.io/projected/9bb0baa8-ce19-4923-9e24-66933db4c618-kube-api-access-jx5bh\") pod \"9bb0baa8-ce19-4923-9e24-66933db4c618\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.072144 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-sb\") pod \"9bb0baa8-ce19-4923-9e24-66933db4c618\" (UID: \"9bb0baa8-ce19-4923-9e24-66933db4c618\") " Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.075263 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-scripts" (OuterVolumeSpecName: "scripts") pod "7350a4c2-fbaf-44d2-afd4-8001bc71b2a4" (UID: "7350a4c2-fbaf-44d2-afd4-8001bc71b2a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.076520 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27a897af-5729-4d7a-94a6-8318aa88eff5-logs" (OuterVolumeSpecName: "logs") pod "27a897af-5729-4d7a-94a6-8318aa88eff5" (UID: "27a897af-5729-4d7a-94a6-8318aa88eff5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.075346 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-config-data" (OuterVolumeSpecName: "config-data") pod "7350a4c2-fbaf-44d2-afd4-8001bc71b2a4" (UID: "7350a4c2-fbaf-44d2-afd4-8001bc71b2a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.076005 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-config-data" (OuterVolumeSpecName: "config-data") pod "27a897af-5729-4d7a-94a6-8318aa88eff5" (UID: "27a897af-5729-4d7a-94a6-8318aa88eff5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.076257 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-scripts" (OuterVolumeSpecName: "scripts") pod "27a897af-5729-4d7a-94a6-8318aa88eff5" (UID: "27a897af-5729-4d7a-94a6-8318aa88eff5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.076755 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-logs" (OuterVolumeSpecName: "logs") pod "7350a4c2-fbaf-44d2-afd4-8001bc71b2a4" (UID: "7350a4c2-fbaf-44d2-afd4-8001bc71b2a4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.104031 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27a897af-5729-4d7a-94a6-8318aa88eff5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "27a897af-5729-4d7a-94a6-8318aa88eff5" (UID: "27a897af-5729-4d7a-94a6-8318aa88eff5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.111263 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-kube-api-access-hv4vq" (OuterVolumeSpecName: "kube-api-access-hv4vq") pod "7350a4c2-fbaf-44d2-afd4-8001bc71b2a4" (UID: "7350a4c2-fbaf-44d2-afd4-8001bc71b2a4"). InnerVolumeSpecName "kube-api-access-hv4vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.111376 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27a897af-5729-4d7a-94a6-8318aa88eff5-kube-api-access-7fbpq" (OuterVolumeSpecName: "kube-api-access-7fbpq") pod "27a897af-5729-4d7a-94a6-8318aa88eff5" (UID: "27a897af-5729-4d7a-94a6-8318aa88eff5"). InnerVolumeSpecName "kube-api-access-7fbpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.113378 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7350a4c2-fbaf-44d2-afd4-8001bc71b2a4" (UID: "7350a4c2-fbaf-44d2-afd4-8001bc71b2a4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.114831 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bb0baa8-ce19-4923-9e24-66933db4c618-kube-api-access-jx5bh" (OuterVolumeSpecName: "kube-api-access-jx5bh") pod "9bb0baa8-ce19-4923-9e24-66933db4c618" (UID: "9bb0baa8-ce19-4923-9e24-66933db4c618"). InnerVolumeSpecName "kube-api-access-jx5bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.146800 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9bb0baa8-ce19-4923-9e24-66933db4c618" (UID: "9bb0baa8-ce19-4923-9e24-66933db4c618"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.148732 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-config" (OuterVolumeSpecName: "config") pod "9bb0baa8-ce19-4923-9e24-66933db4c618" (UID: "9bb0baa8-ce19-4923-9e24-66933db4c618"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.170990 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9bb0baa8-ce19-4923-9e24-66933db4c618" (UID: "9bb0baa8-ce19-4923-9e24-66933db4c618"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175673 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx5bh\" (UniqueName: \"kubernetes.io/projected/9bb0baa8-ce19-4923-9e24-66933db4c618-kube-api-access-jx5bh\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175702 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175712 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175719 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175727 4719 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/27a897af-5729-4d7a-94a6-8318aa88eff5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175735 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175774 4719 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175784 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175792 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fbpq\" (UniqueName: \"kubernetes.io/projected/27a897af-5729-4d7a-94a6-8318aa88eff5-kube-api-access-7fbpq\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175800 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175808 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv4vq\" (UniqueName: \"kubernetes.io/projected/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-kube-api-access-hv4vq\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175817 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27a897af-5729-4d7a-94a6-8318aa88eff5-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175825 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27a897af-5729-4d7a-94a6-8318aa88eff5-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.175832 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.181237 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9bb0baa8-ce19-4923-9e24-66933db4c618" (UID: "9bb0baa8-ce19-4923-9e24-66933db4c618"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.277586 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bb0baa8-ce19-4923-9e24-66933db4c618-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.304194 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5dc4d6cbb7-scwpt" event={"ID":"27a897af-5729-4d7a-94a6-8318aa88eff5","Type":"ContainerDied","Data":"9eb109d8aa3ce8963336a2282087fadd7ff10a6841a607234cfb123bd3481113"} Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.304298 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dc4d6cbb7-scwpt" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.312966 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" event={"ID":"9bb0baa8-ce19-4923-9e24-66933db4c618","Type":"ContainerDied","Data":"2b8b726944c1c2880bff5013ca1db9803be59ad45707d31ccffb449d2bcf64be"} Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.313055 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.337788 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-677f6b8777-tsxng" event={"ID":"16a1675a-ef86-4636-bb06-dfd592019a53","Type":"ContainerDied","Data":"bf5685a8b7ce2f086b8b97d7a354e0c15af7570b634054a3cf21951c20f885d2"} Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.337951 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-677f6b8777-tsxng" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.344213 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f46b45545-rn64s" event={"ID":"7350a4c2-fbaf-44d2-afd4-8001bc71b2a4","Type":"ContainerDied","Data":"0ae2c78d22e6fa1faf94da7f965488dc33332029ac10a20f1f7f0a1ee39a97a9"} Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.344488 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f46b45545-rn64s" Dec 15 12:38:21 crc kubenswrapper[4719]: E1215 12:38:21.360324 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-8qr2l" podUID="58fea5e6-9ff9-4a44-be8a-311735af3e4b" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.461503 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5dc4d6cbb7-scwpt"] Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.470696 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5dc4d6cbb7-scwpt"] Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.502725 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27a897af-5729-4d7a-94a6-8318aa88eff5" path="/var/lib/kubelet/pods/27a897af-5729-4d7a-94a6-8318aa88eff5/volumes" Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.503228 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-m9j55"] Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.505048 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-m9j55"] Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.518097 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-677f6b8777-tsxng"] Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.524722 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bc5fb48-jqf6b"] Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.532984 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-677f6b8777-tsxng"] Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.547993 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f46b45545-rn64s"] Dec 15 12:38:21 crc kubenswrapper[4719]: I1215 12:38:21.556586 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f46b45545-rn64s"] Dec 15 12:38:22 crc kubenswrapper[4719]: W1215 12:38:22.628953 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9b7dc15_da8d_41ad_a5ab_cb7e16b0c926.slice/crio-13edec4bd52d061c8f7b47ec73409016af911bff76f205c60311104c2d300363 WatchSource:0}: Error finding container 13edec4bd52d061c8f7b47ec73409016af911bff76f205c60311104c2d300363: Status 404 returned error can't find the container with id 13edec4bd52d061c8f7b47ec73409016af911bff76f205c60311104c2d300363 Dec 15 12:38:22 crc kubenswrapper[4719]: I1215 12:38:22.649567 4719 scope.go:117] "RemoveContainer" containerID="103948537bba25ac67d5771374cc139f4b1d0eceeb003de13c0219aae21ecb6c" Dec 15 12:38:22 crc kubenswrapper[4719]: E1215 12:38:22.652886 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 15 12:38:22 crc kubenswrapper[4719]: E1215 12:38:22.653047 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4drnb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-gprbv_openstack(ef9ed6e9-3050-4335-8ffc-97f889eb63e6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:38:22 crc kubenswrapper[4719]: E1215 12:38:22.654483 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-gprbv" podUID="ef9ed6e9-3050-4335-8ffc-97f889eb63e6" Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.162039 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c6bcf58b6-vx99l"] Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.323181 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.366065 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc5fb48-jqf6b" event={"ID":"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926","Type":"ContainerStarted","Data":"13edec4bd52d061c8f7b47ec73409016af911bff76f205c60311104c2d300363"} Dec 15 12:38:23 crc kubenswrapper[4719]: E1215 12:38:23.367434 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-gprbv" podUID="ef9ed6e9-3050-4335-8ffc-97f889eb63e6" Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.390687 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-m9j55" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 15 12:38:23 crc kubenswrapper[4719]: E1215 12:38:23.476014 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified" Dec 15 12:38:23 crc kubenswrapper[4719]: E1215 12:38:23.476210 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-notification-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8dh578hdfhb6h686h55ch54ch648hb8h585h547h78h56chb9h5d4h686h6fhcdh64dh84hc4h5ch95h8ch56dh54dh5bh5fh66ch58bhchd6q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-notification-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rmmgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/notificationhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(edd905bf-6348-446d-abe8-fb632760bc1f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.490001 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16a1675a-ef86-4636-bb06-dfd592019a53" path="/var/lib/kubelet/pods/16a1675a-ef86-4636-bb06-dfd592019a53/volumes" Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.490451 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7350a4c2-fbaf-44d2-afd4-8001bc71b2a4" path="/var/lib/kubelet/pods/7350a4c2-fbaf-44d2-afd4-8001bc71b2a4/volumes" Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.490836 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" path="/var/lib/kubelet/pods/9bb0baa8-ce19-4923-9e24-66933db4c618/volumes" Dec 15 12:38:23 crc kubenswrapper[4719]: W1215 12:38:23.532441 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51211ffc_0363_4e48_9972_15b8e997d4de.slice/crio-b89c621b3f445671a6352d562d6278fc19f7474fd4d123453295abaaf25e9f8f WatchSource:0}: Error finding container b89c621b3f445671a6352d562d6278fc19f7474fd4d123453295abaaf25e9f8f: Status 404 returned error can't find the container with id b89c621b3f445671a6352d562d6278fc19f7474fd4d123453295abaaf25e9f8f Dec 15 12:38:23 crc kubenswrapper[4719]: W1215 12:38:23.535903 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae2ae4f5_b3a6_4f26_b897_2dc0cb588f6e.slice/crio-228c4f698c30ff9eb50ccbb80912dae089a8469028a3ce638228519ede1ea32b WatchSource:0}: Error finding container 228c4f698c30ff9eb50ccbb80912dae089a8469028a3ce638228519ede1ea32b: Status 404 returned error can't find the container with id 228c4f698c30ff9eb50ccbb80912dae089a8469028a3ce638228519ede1ea32b Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.572226 4719 scope.go:117] "RemoveContainer" containerID="93daf2ea66ab371ecf3c3b31844e8f1700c620068a05444febad6f71febba7da" Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.610444 4719 scope.go:117] "RemoveContainer" containerID="0e96ac6e3ea981f737bec1c9446e922db723f7ab9152cce21ff8a3f7e3d80063" Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.764568 4719 scope.go:117] "RemoveContainer" containerID="29cbf0603161715ba41954c2791f0169f0df10391331aac7a7dae583f4bd69af" Dec 15 12:38:23 crc kubenswrapper[4719]: I1215 12:38:23.996212 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-88rcn"] Dec 15 12:38:24 crc kubenswrapper[4719]: W1215 12:38:24.012088 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod978132fa_fca3_4665_b133_682da5d02f93.slice/crio-e40ff5c6451be44b859ffa6c63b221381e2efaf497627d79bc9c2a7286622c57 WatchSource:0}: Error finding container e40ff5c6451be44b859ffa6c63b221381e2efaf497627d79bc9c2a7286622c57: Status 404 returned error can't find the container with id e40ff5c6451be44b859ffa6c63b221381e2efaf497627d79bc9c2a7286622c57 Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.018007 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.149460 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.378058 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6bcf58b6-vx99l" event={"ID":"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e","Type":"ContainerStarted","Data":"228c4f698c30ff9eb50ccbb80912dae089a8469028a3ce638228519ede1ea32b"} Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.382921 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63366e84-551e-4eb0-a007-7156b6ab1d6b","Type":"ContainerStarted","Data":"580493cf67a7c6217cf5b3d1230cd2f7eb90d067ae3c56b2533df1dad3da25f3"} Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.399462 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-88rcn" event={"ID":"978132fa-fca3-4665-b133-682da5d02f93","Type":"ContainerStarted","Data":"f2b161a88efddcd95d6d9476aa89c6ab52a625caa07795376362d6a570e0cdf2"} Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.399504 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-88rcn" event={"ID":"978132fa-fca3-4665-b133-682da5d02f93","Type":"ContainerStarted","Data":"e40ff5c6451be44b859ffa6c63b221381e2efaf497627d79bc9c2a7286622c57"} Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.443403 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51211ffc-0363-4e48-9972-15b8e997d4de","Type":"ContainerStarted","Data":"2f5fc5d9ab5592e8a600bbf12061874b624c5e60975cf97b442cd6354eacc609"} Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.443437 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-88rcn" podStartSLOduration=17.443424759 podStartE2EDuration="17.443424759s" podCreationTimestamp="2025-12-15 12:38:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:38:24.431189838 +0000 UTC m=+1265.373482878" watchObservedRunningTime="2025-12-15 12:38:24.443424759 +0000 UTC m=+1265.385717799" Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.443476 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51211ffc-0363-4e48-9972-15b8e997d4de","Type":"ContainerStarted","Data":"b89c621b3f445671a6352d562d6278fc19f7474fd4d123453295abaaf25e9f8f"} Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.451382 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nd852" event={"ID":"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c","Type":"ContainerStarted","Data":"f78a6395ccee6de4c3c5dec602c3cb6643a39d904cb1023542b07be5867f3e20"} Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.458196 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc5fb48-jqf6b" event={"ID":"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926","Type":"ContainerStarted","Data":"975b276955c100d7fd0352b04978b9b7d4a57b2a3ddba6e0cbb4fd095434fee3"} Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.458240 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc5fb48-jqf6b" event={"ID":"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926","Type":"ContainerStarted","Data":"4f189d10fd778489882ae3909af6d12fd08375b585a24ca4e01f66ff70e4b5ff"} Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.475408 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nd852" podStartSLOduration=4.033169373 podStartE2EDuration="47.475391264s" podCreationTimestamp="2025-12-15 12:37:37 +0000 UTC" firstStartedPulling="2025-12-15 12:37:40.20217206 +0000 UTC m=+1221.144465090" lastFinishedPulling="2025-12-15 12:38:23.644393951 +0000 UTC m=+1264.586686981" observedRunningTime="2025-12-15 12:38:24.471225314 +0000 UTC m=+1265.413518354" watchObservedRunningTime="2025-12-15 12:38:24.475391264 +0000 UTC m=+1265.417684294" Dec 15 12:38:24 crc kubenswrapper[4719]: I1215 12:38:24.516534 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6bc5fb48-jqf6b" podStartSLOduration=36.387507623 podStartE2EDuration="37.516509462s" podCreationTimestamp="2025-12-15 12:37:47 +0000 UTC" firstStartedPulling="2025-12-15 12:38:22.637626084 +0000 UTC m=+1263.579919114" lastFinishedPulling="2025-12-15 12:38:23.766627923 +0000 UTC m=+1264.708920953" observedRunningTime="2025-12-15 12:38:24.49487498 +0000 UTC m=+1265.437168030" watchObservedRunningTime="2025-12-15 12:38:24.516509462 +0000 UTC m=+1265.458802492" Dec 15 12:38:25 crc kubenswrapper[4719]: I1215 12:38:25.501742 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51211ffc-0363-4e48-9972-15b8e997d4de","Type":"ContainerStarted","Data":"04f9a53ed870a660d04118a7bc49f93fd4196e4150a668f6be6bcc5b0b375c32"} Dec 15 12:38:25 crc kubenswrapper[4719]: I1215 12:38:25.516362 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6bcf58b6-vx99l" event={"ID":"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e","Type":"ContainerStarted","Data":"a515a01ab9e93d2a61d2349f044b8132b8de12998c6a016d169b104d98e71a08"} Dec 15 12:38:25 crc kubenswrapper[4719]: I1215 12:38:25.516410 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6bcf58b6-vx99l" event={"ID":"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e","Type":"ContainerStarted","Data":"c255291bba1cc5f6daef1060c061c10c9363cee215f9bd734f39e210ffbe638b"} Dec 15 12:38:25 crc kubenswrapper[4719]: I1215 12:38:25.518539 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63366e84-551e-4eb0-a007-7156b6ab1d6b","Type":"ContainerStarted","Data":"ac4e16b1d34616e8a23c7400e0474377f48d038a2b38a39b0094520b9e4d1a07"} Dec 15 12:38:25 crc kubenswrapper[4719]: I1215 12:38:25.538448 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=33.538432261 podStartE2EDuration="33.538432261s" podCreationTimestamp="2025-12-15 12:37:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:38:25.532117904 +0000 UTC m=+1266.474410934" watchObservedRunningTime="2025-12-15 12:38:25.538432261 +0000 UTC m=+1266.480725291" Dec 15 12:38:25 crc kubenswrapper[4719]: I1215 12:38:25.579474 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c6bcf58b6-vx99l" podStartSLOduration=38.607106849 podStartE2EDuration="39.579447076s" podCreationTimestamp="2025-12-15 12:37:46 +0000 UTC" firstStartedPulling="2025-12-15 12:38:23.539207721 +0000 UTC m=+1264.481500751" lastFinishedPulling="2025-12-15 12:38:24.511547948 +0000 UTC m=+1265.453840978" observedRunningTime="2025-12-15 12:38:25.574377388 +0000 UTC m=+1266.516670418" watchObservedRunningTime="2025-12-15 12:38:25.579447076 +0000 UTC m=+1266.521740106" Dec 15 12:38:26 crc kubenswrapper[4719]: I1215 12:38:26.538734 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63366e84-551e-4eb0-a007-7156b6ab1d6b","Type":"ContainerStarted","Data":"be535f9264accf5731790a1bfd0b7a52d63b8830d8427c919804c75fedbd7290"} Dec 15 12:38:26 crc kubenswrapper[4719]: I1215 12:38:26.573368 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=39.573346943 podStartE2EDuration="39.573346943s" podCreationTimestamp="2025-12-15 12:37:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:38:26.563113325 +0000 UTC m=+1267.505406365" watchObservedRunningTime="2025-12-15 12:38:26.573346943 +0000 UTC m=+1267.515639973" Dec 15 12:38:27 crc kubenswrapper[4719]: I1215 12:38:27.389302 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:38:27 crc kubenswrapper[4719]: I1215 12:38:27.389630 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:38:27 crc kubenswrapper[4719]: I1215 12:38:27.574421 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:38:27 crc kubenswrapper[4719]: I1215 12:38:27.574476 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:38:27 crc kubenswrapper[4719]: I1215 12:38:27.598266 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 15 12:38:27 crc kubenswrapper[4719]: I1215 12:38:27.598318 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 15 12:38:27 crc kubenswrapper[4719]: I1215 12:38:27.649452 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 15 12:38:27 crc kubenswrapper[4719]: I1215 12:38:27.650748 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 15 12:38:28 crc kubenswrapper[4719]: I1215 12:38:28.557749 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 15 12:38:28 crc kubenswrapper[4719]: I1215 12:38:28.557805 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 15 12:38:31 crc kubenswrapper[4719]: I1215 12:38:31.590872 4719 generic.go:334] "Generic (PLEG): container finished" podID="978132fa-fca3-4665-b133-682da5d02f93" containerID="f2b161a88efddcd95d6d9476aa89c6ab52a625caa07795376362d6a570e0cdf2" exitCode=0 Dec 15 12:38:31 crc kubenswrapper[4719]: I1215 12:38:31.590955 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-88rcn" event={"ID":"978132fa-fca3-4665-b133-682da5d02f93","Type":"ContainerDied","Data":"f2b161a88efddcd95d6d9476aa89c6ab52a625caa07795376362d6a570e0cdf2"} Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.146681 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.146722 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.210726 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.232159 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.609222 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.612877 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-88rcn" event={"ID":"978132fa-fca3-4665-b133-682da5d02f93","Type":"ContainerDied","Data":"e40ff5c6451be44b859ffa6c63b221381e2efaf497627d79bc9c2a7286622c57"} Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.612924 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e40ff5c6451be44b859ffa6c63b221381e2efaf497627d79bc9c2a7286622c57" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.615545 4719 generic.go:334] "Generic (PLEG): container finished" podID="cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" containerID="f78a6395ccee6de4c3c5dec602c3cb6643a39d904cb1023542b07be5867f3e20" exitCode=0 Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.615633 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nd852" event={"ID":"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c","Type":"ContainerDied","Data":"f78a6395ccee6de4c3c5dec602c3cb6643a39d904cb1023542b07be5867f3e20"} Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.616327 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.616349 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.770400 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pzzj\" (UniqueName: \"kubernetes.io/projected/978132fa-fca3-4665-b133-682da5d02f93-kube-api-access-4pzzj\") pod \"978132fa-fca3-4665-b133-682da5d02f93\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.770455 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-config-data\") pod \"978132fa-fca3-4665-b133-682da5d02f93\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.770609 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-credential-keys\") pod \"978132fa-fca3-4665-b133-682da5d02f93\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.770658 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-scripts\") pod \"978132fa-fca3-4665-b133-682da5d02f93\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.770687 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-fernet-keys\") pod \"978132fa-fca3-4665-b133-682da5d02f93\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.770739 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-combined-ca-bundle\") pod \"978132fa-fca3-4665-b133-682da5d02f93\" (UID: \"978132fa-fca3-4665-b133-682da5d02f93\") " Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.776306 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-scripts" (OuterVolumeSpecName: "scripts") pod "978132fa-fca3-4665-b133-682da5d02f93" (UID: "978132fa-fca3-4665-b133-682da5d02f93"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.776688 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "978132fa-fca3-4665-b133-682da5d02f93" (UID: "978132fa-fca3-4665-b133-682da5d02f93"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.779077 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/978132fa-fca3-4665-b133-682da5d02f93-kube-api-access-4pzzj" (OuterVolumeSpecName: "kube-api-access-4pzzj") pod "978132fa-fca3-4665-b133-682da5d02f93" (UID: "978132fa-fca3-4665-b133-682da5d02f93"). InnerVolumeSpecName "kube-api-access-4pzzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.779174 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "978132fa-fca3-4665-b133-682da5d02f93" (UID: "978132fa-fca3-4665-b133-682da5d02f93"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.807275 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "978132fa-fca3-4665-b133-682da5d02f93" (UID: "978132fa-fca3-4665-b133-682da5d02f93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.820424 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-config-data" (OuterVolumeSpecName: "config-data") pod "978132fa-fca3-4665-b133-682da5d02f93" (UID: "978132fa-fca3-4665-b133-682da5d02f93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.873816 4719 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.873847 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.873871 4719 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.873880 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.873888 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pzzj\" (UniqueName: \"kubernetes.io/projected/978132fa-fca3-4665-b133-682da5d02f93-kube-api-access-4pzzj\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:33 crc kubenswrapper[4719]: I1215 12:38:33.873899 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978132fa-fca3-4665-b133-682da5d02f93-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.624481 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edd905bf-6348-446d-abe8-fb632760bc1f","Type":"ContainerStarted","Data":"6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28"} Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.624568 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-88rcn" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.800992 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5c7789bfff-vvvqf"] Dec 15 12:38:34 crc kubenswrapper[4719]: E1215 12:38:34.801827 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="978132fa-fca3-4665-b133-682da5d02f93" containerName="keystone-bootstrap" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.801844 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="978132fa-fca3-4665-b133-682da5d02f93" containerName="keystone-bootstrap" Dec 15 12:38:34 crc kubenswrapper[4719]: E1215 12:38:34.801952 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="init" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.801962 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="init" Dec 15 12:38:34 crc kubenswrapper[4719]: E1215 12:38:34.801980 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="dnsmasq-dns" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.801988 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="dnsmasq-dns" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.802218 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="978132fa-fca3-4665-b133-682da5d02f93" containerName="keystone-bootstrap" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.802242 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bb0baa8-ce19-4923-9e24-66933db4c618" containerName="dnsmasq-dns" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.802945 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.809228 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.809379 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.809462 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.809646 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.812034 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.812044 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kdx8k" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.848417 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c7789bfff-vvvqf"] Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.907620 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgs6k\" (UniqueName: \"kubernetes.io/projected/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-kube-api-access-vgs6k\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.908995 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-config-data\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.909091 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-credential-keys\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.909272 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-combined-ca-bundle\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.909356 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-internal-tls-certs\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.909451 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-scripts\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.909540 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-public-tls-certs\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:34 crc kubenswrapper[4719]: I1215 12:38:34.909656 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-fernet-keys\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.012580 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-combined-ca-bundle\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.012618 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-internal-tls-certs\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.012644 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-scripts\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.012679 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-public-tls-certs\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.012710 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-fernet-keys\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.012737 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgs6k\" (UniqueName: \"kubernetes.io/projected/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-kube-api-access-vgs6k\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.012759 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-config-data\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.012779 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-credential-keys\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.027566 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-combined-ca-bundle\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.033686 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-fernet-keys\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.033933 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-internal-tls-certs\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.034274 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-public-tls-certs\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.034670 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-credential-keys\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.034728 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-config-data\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.043378 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-scripts\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.049321 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgs6k\" (UniqueName: \"kubernetes.io/projected/e087fc68-a428-4d63-b1c1-9d9a6b7b0c93-kube-api-access-vgs6k\") pod \"keystone-5c7789bfff-vvvqf\" (UID: \"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93\") " pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.114315 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nd852" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.133789 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.214611 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-config-data\") pod \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.214654 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-logs\") pod \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.214789 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlds7\" (UniqueName: \"kubernetes.io/projected/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-kube-api-access-rlds7\") pod \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.214890 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-combined-ca-bundle\") pod \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.214941 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-scripts\") pod \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\" (UID: \"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c\") " Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.219939 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-logs" (OuterVolumeSpecName: "logs") pod "cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" (UID: "cba0d08d-bba6-4a68-bc44-c4cf1619ad4c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.225149 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-scripts" (OuterVolumeSpecName: "scripts") pod "cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" (UID: "cba0d08d-bba6-4a68-bc44-c4cf1619ad4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.241809 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-kube-api-access-rlds7" (OuterVolumeSpecName: "kube-api-access-rlds7") pod "cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" (UID: "cba0d08d-bba6-4a68-bc44-c4cf1619ad4c"). InnerVolumeSpecName "kube-api-access-rlds7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.246963 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" (UID: "cba0d08d-bba6-4a68-bc44-c4cf1619ad4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.257714 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-config-data" (OuterVolumeSpecName: "config-data") pod "cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" (UID: "cba0d08d-bba6-4a68-bc44-c4cf1619ad4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.317261 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.317318 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.317333 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlds7\" (UniqueName: \"kubernetes.io/projected/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-kube-api-access-rlds7\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.317347 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.317357 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.641420 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8qr2l" event={"ID":"58fea5e6-9ff9-4a44-be8a-311735af3e4b","Type":"ContainerStarted","Data":"700804ea29d99bd4f42fffec72e1606580a670faee05531a7f62112e644c7a84"} Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.658210 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nd852" event={"ID":"cba0d08d-bba6-4a68-bc44-c4cf1619ad4c","Type":"ContainerDied","Data":"b79d28b6888231612f46ec33c6ef828cd2656d4a0c24922933c5ac29e440d166"} Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.658251 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b79d28b6888231612f46ec33c6ef828cd2656d4a0c24922933c5ac29e440d166" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.658311 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nd852" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.675699 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-8qr2l" podStartSLOduration=3.5453273899999997 podStartE2EDuration="58.675678248s" podCreationTimestamp="2025-12-15 12:37:37 +0000 UTC" firstStartedPulling="2025-12-15 12:37:40.091831018 +0000 UTC m=+1221.034124038" lastFinishedPulling="2025-12-15 12:38:35.222181856 +0000 UTC m=+1276.164474896" observedRunningTime="2025-12-15 12:38:35.668746182 +0000 UTC m=+1276.611039212" watchObservedRunningTime="2025-12-15 12:38:35.675678248 +0000 UTC m=+1276.617971278" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.754577 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c7789bfff-vvvqf"] Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.796793 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9d9f8bcf4-hj8fd"] Dec 15 12:38:35 crc kubenswrapper[4719]: W1215 12:38:35.799460 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode087fc68_a428_4d63_b1c1_9d9a6b7b0c93.slice/crio-4f88ad6f816aa320f48fb32932c4e1a9e46ac59a27b17ab1ebe1dda474fad840 WatchSource:0}: Error finding container 4f88ad6f816aa320f48fb32932c4e1a9e46ac59a27b17ab1ebe1dda474fad840: Status 404 returned error can't find the container with id 4f88ad6f816aa320f48fb32932c4e1a9e46ac59a27b17ab1ebe1dda474fad840 Dec 15 12:38:35 crc kubenswrapper[4719]: E1215 12:38:35.799712 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" containerName="placement-db-sync" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.799784 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" containerName="placement-db-sync" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.800005 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" containerName="placement-db-sync" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.800922 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.805481 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.806052 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vbw9r" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.806194 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.806248 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.806408 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.836310 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9d9f8bcf4-hj8fd"] Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.938250 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-public-tls-certs\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.938307 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-combined-ca-bundle\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.938328 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac288046-bbc8-4521-84b8-c16e7d096217-logs\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.938382 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-internal-tls-certs\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.938412 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx2lf\" (UniqueName: \"kubernetes.io/projected/ac288046-bbc8-4521-84b8-c16e7d096217-kube-api-access-mx2lf\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.938431 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-scripts\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:35 crc kubenswrapper[4719]: I1215 12:38:35.938466 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-config-data\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.041760 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-scripts\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.042250 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-config-data\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.042295 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-public-tls-certs\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.042345 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-combined-ca-bundle\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.042364 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac288046-bbc8-4521-84b8-c16e7d096217-logs\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.042423 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-internal-tls-certs\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.042461 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx2lf\" (UniqueName: \"kubernetes.io/projected/ac288046-bbc8-4521-84b8-c16e7d096217-kube-api-access-mx2lf\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.048720 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac288046-bbc8-4521-84b8-c16e7d096217-logs\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.050836 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-combined-ca-bundle\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.052703 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-scripts\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.052728 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-public-tls-certs\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.053207 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-config-data\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.056433 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac288046-bbc8-4521-84b8-c16e7d096217-internal-tls-certs\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.068536 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx2lf\" (UniqueName: \"kubernetes.io/projected/ac288046-bbc8-4521-84b8-c16e7d096217-kube-api-access-mx2lf\") pod \"placement-9d9f8bcf4-hj8fd\" (UID: \"ac288046-bbc8-4521-84b8-c16e7d096217\") " pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.136608 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.597727 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9d9f8bcf4-hj8fd"] Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.684051 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9d9f8bcf4-hj8fd" event={"ID":"ac288046-bbc8-4521-84b8-c16e7d096217","Type":"ContainerStarted","Data":"01a89d94ebec3e8c690e8d5b4ee40c3936c43ee0abc7796467c7dbbe8fd7ee0e"} Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.685176 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c7789bfff-vvvqf" event={"ID":"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93","Type":"ContainerStarted","Data":"66fd0d6132e4c19462e3d5a72028862f73d256bba34ca944900856e7c6781472"} Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.685424 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c7789bfff-vvvqf" event={"ID":"e087fc68-a428-4d63-b1c1-9d9a6b7b0c93","Type":"ContainerStarted","Data":"4f88ad6f816aa320f48fb32932c4e1a9e46ac59a27b17ab1ebe1dda474fad840"} Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.687139 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:38:36 crc kubenswrapper[4719]: I1215 12:38:36.729695 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5c7789bfff-vvvqf" podStartSLOduration=2.729673694 podStartE2EDuration="2.729673694s" podCreationTimestamp="2025-12-15 12:38:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:38:36.716321159 +0000 UTC m=+1277.658614189" watchObservedRunningTime="2025-12-15 12:38:36.729673694 +0000 UTC m=+1277.671966724" Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.181173 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.231078 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.392924 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.576645 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc5fb48-jqf6b" podUID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.730560 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9d9f8bcf4-hj8fd" event={"ID":"ac288046-bbc8-4521-84b8-c16e7d096217","Type":"ContainerStarted","Data":"f6e6029c12c5775afb19d67ed0e32555cab1074147e9ddea94b916a2554f1a91"} Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.730596 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9d9f8bcf4-hj8fd" event={"ID":"ac288046-bbc8-4521-84b8-c16e7d096217","Type":"ContainerStarted","Data":"8037d5837d350adcb9d8a2b28278ad859883d0562f8d57ef79f1952f9c7b5656"} Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.731061 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.731083 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.733057 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.733179 4719 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.759348 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-9d9f8bcf4-hj8fd" podStartSLOduration=2.759334334 podStartE2EDuration="2.759334334s" podCreationTimestamp="2025-12-15 12:38:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:38:37.759033615 +0000 UTC m=+1278.701326645" watchObservedRunningTime="2025-12-15 12:38:37.759334334 +0000 UTC m=+1278.701627364" Dec 15 12:38:37 crc kubenswrapper[4719]: I1215 12:38:37.763730 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 15 12:38:38 crc kubenswrapper[4719]: I1215 12:38:38.743077 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gprbv" event={"ID":"ef9ed6e9-3050-4335-8ffc-97f889eb63e6","Type":"ContainerStarted","Data":"68928debcaf4ded9eda72c222dfdb8418b8bf08b43ca2b04dfd19ea947b50a35"} Dec 15 12:38:38 crc kubenswrapper[4719]: I1215 12:38:38.767264 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-gprbv" podStartSLOduration=4.437248558 podStartE2EDuration="1m1.767242667s" podCreationTimestamp="2025-12-15 12:37:37 +0000 UTC" firstStartedPulling="2025-12-15 12:37:39.845018794 +0000 UTC m=+1220.787311824" lastFinishedPulling="2025-12-15 12:38:37.175012903 +0000 UTC m=+1278.117305933" observedRunningTime="2025-12-15 12:38:38.764153701 +0000 UTC m=+1279.706446731" watchObservedRunningTime="2025-12-15 12:38:38.767242667 +0000 UTC m=+1279.709535707" Dec 15 12:38:44 crc kubenswrapper[4719]: I1215 12:38:44.797503 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8qr2l" event={"ID":"58fea5e6-9ff9-4a44-be8a-311735af3e4b","Type":"ContainerDied","Data":"700804ea29d99bd4f42fffec72e1606580a670faee05531a7f62112e644c7a84"} Dec 15 12:38:44 crc kubenswrapper[4719]: I1215 12:38:44.797468 4719 generic.go:334] "Generic (PLEG): container finished" podID="58fea5e6-9ff9-4a44-be8a-311735af3e4b" containerID="700804ea29d99bd4f42fffec72e1606580a670faee05531a7f62112e644c7a84" exitCode=0 Dec 15 12:38:45 crc kubenswrapper[4719]: E1215 12:38:45.473392 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"ceilometer-notification-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="edd905bf-6348-446d-abe8-fb632760bc1f" Dec 15 12:38:45 crc kubenswrapper[4719]: I1215 12:38:45.835943 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="edd905bf-6348-446d-abe8-fb632760bc1f" containerName="sg-core" containerID="cri-o://6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28" gracePeriod=30 Dec 15 12:38:45 crc kubenswrapper[4719]: I1215 12:38:45.836354 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="edd905bf-6348-446d-abe8-fb632760bc1f" containerName="proxy-httpd" containerID="cri-o://2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558" gracePeriod=30 Dec 15 12:38:45 crc kubenswrapper[4719]: I1215 12:38:45.836178 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edd905bf-6348-446d-abe8-fb632760bc1f","Type":"ContainerStarted","Data":"2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558"} Dec 15 12:38:45 crc kubenswrapper[4719]: I1215 12:38:45.836612 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.181685 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.254461 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxzg5\" (UniqueName: \"kubernetes.io/projected/58fea5e6-9ff9-4a44-be8a-311735af3e4b-kube-api-access-fxzg5\") pod \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.254824 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-db-sync-config-data\") pod \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.254903 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-combined-ca-bundle\") pod \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\" (UID: \"58fea5e6-9ff9-4a44-be8a-311735af3e4b\") " Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.261486 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58fea5e6-9ff9-4a44-be8a-311735af3e4b-kube-api-access-fxzg5" (OuterVolumeSpecName: "kube-api-access-fxzg5") pod "58fea5e6-9ff9-4a44-be8a-311735af3e4b" (UID: "58fea5e6-9ff9-4a44-be8a-311735af3e4b"). InnerVolumeSpecName "kube-api-access-fxzg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.262594 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "58fea5e6-9ff9-4a44-be8a-311735af3e4b" (UID: "58fea5e6-9ff9-4a44-be8a-311735af3e4b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.342025 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58fea5e6-9ff9-4a44-be8a-311735af3e4b" (UID: "58fea5e6-9ff9-4a44-be8a-311735af3e4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.358515 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxzg5\" (UniqueName: \"kubernetes.io/projected/58fea5e6-9ff9-4a44-be8a-311735af3e4b-kube-api-access-fxzg5\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.358555 4719 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.358564 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58fea5e6-9ff9-4a44-be8a-311735af3e4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.462203 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.560589 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-config-data\") pod \"edd905bf-6348-446d-abe8-fb632760bc1f\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.560747 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-combined-ca-bundle\") pod \"edd905bf-6348-446d-abe8-fb632760bc1f\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.560811 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmmgs\" (UniqueName: \"kubernetes.io/projected/edd905bf-6348-446d-abe8-fb632760bc1f-kube-api-access-rmmgs\") pod \"edd905bf-6348-446d-abe8-fb632760bc1f\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.560837 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-run-httpd\") pod \"edd905bf-6348-446d-abe8-fb632760bc1f\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.560950 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-scripts\") pod \"edd905bf-6348-446d-abe8-fb632760bc1f\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.561004 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-log-httpd\") pod \"edd905bf-6348-446d-abe8-fb632760bc1f\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.561038 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-sg-core-conf-yaml\") pod \"edd905bf-6348-446d-abe8-fb632760bc1f\" (UID: \"edd905bf-6348-446d-abe8-fb632760bc1f\") " Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.561326 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "edd905bf-6348-446d-abe8-fb632760bc1f" (UID: "edd905bf-6348-446d-abe8-fb632760bc1f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.561440 4719 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.561900 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "edd905bf-6348-446d-abe8-fb632760bc1f" (UID: "edd905bf-6348-446d-abe8-fb632760bc1f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.566751 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-scripts" (OuterVolumeSpecName: "scripts") pod "edd905bf-6348-446d-abe8-fb632760bc1f" (UID: "edd905bf-6348-446d-abe8-fb632760bc1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.571078 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edd905bf-6348-446d-abe8-fb632760bc1f-kube-api-access-rmmgs" (OuterVolumeSpecName: "kube-api-access-rmmgs") pod "edd905bf-6348-446d-abe8-fb632760bc1f" (UID: "edd905bf-6348-446d-abe8-fb632760bc1f"). InnerVolumeSpecName "kube-api-access-rmmgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.592017 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "edd905bf-6348-446d-abe8-fb632760bc1f" (UID: "edd905bf-6348-446d-abe8-fb632760bc1f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.603014 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edd905bf-6348-446d-abe8-fb632760bc1f" (UID: "edd905bf-6348-446d-abe8-fb632760bc1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.623213 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-config-data" (OuterVolumeSpecName: "config-data") pod "edd905bf-6348-446d-abe8-fb632760bc1f" (UID: "edd905bf-6348-446d-abe8-fb632760bc1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.663374 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.663416 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.663432 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmmgs\" (UniqueName: \"kubernetes.io/projected/edd905bf-6348-446d-abe8-fb632760bc1f-kube-api-access-rmmgs\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.663443 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.663474 4719 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edd905bf-6348-446d-abe8-fb632760bc1f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.663483 4719 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/edd905bf-6348-446d-abe8-fb632760bc1f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.844640 4719 generic.go:334] "Generic (PLEG): container finished" podID="edd905bf-6348-446d-abe8-fb632760bc1f" containerID="2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558" exitCode=0 Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.844666 4719 generic.go:334] "Generic (PLEG): container finished" podID="edd905bf-6348-446d-abe8-fb632760bc1f" containerID="6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28" exitCode=2 Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.844714 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edd905bf-6348-446d-abe8-fb632760bc1f","Type":"ContainerDied","Data":"2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558"} Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.844741 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edd905bf-6348-446d-abe8-fb632760bc1f","Type":"ContainerDied","Data":"6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28"} Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.844751 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edd905bf-6348-446d-abe8-fb632760bc1f","Type":"ContainerDied","Data":"67f2f74f3a0bc116fef224d4f19e8313a697d86c2fc7d28016df2f048abdd183"} Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.844765 4719 scope.go:117] "RemoveContainer" containerID="2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.844714 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.847526 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8qr2l" event={"ID":"58fea5e6-9ff9-4a44-be8a-311735af3e4b","Type":"ContainerDied","Data":"698165fdb097b663725a4116e1b32096409d5f48c974b38153507360e8bd7930"} Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.847559 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="698165fdb097b663725a4116e1b32096409d5f48c974b38153507360e8bd7930" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.847559 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8qr2l" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.849225 4719 generic.go:334] "Generic (PLEG): container finished" podID="3243ab84-0a4c-4c74-83de-693c1c96e89e" containerID="f9a10627eba3e330e4bc0909a38d7d0e9a3da28f30c8c1e92e26e0c385194bd6" exitCode=0 Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.849268 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9vwxh" event={"ID":"3243ab84-0a4c-4c74-83de-693c1c96e89e","Type":"ContainerDied","Data":"f9a10627eba3e330e4bc0909a38d7d0e9a3da28f30c8c1e92e26e0c385194bd6"} Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.863497 4719 scope.go:117] "RemoveContainer" containerID="6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.887111 4719 scope.go:117] "RemoveContainer" containerID="2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558" Dec 15 12:38:46 crc kubenswrapper[4719]: E1215 12:38:46.887469 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558\": container with ID starting with 2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558 not found: ID does not exist" containerID="2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.887499 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558"} err="failed to get container status \"2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558\": rpc error: code = NotFound desc = could not find container \"2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558\": container with ID starting with 2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558 not found: ID does not exist" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.887519 4719 scope.go:117] "RemoveContainer" containerID="6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28" Dec 15 12:38:46 crc kubenswrapper[4719]: E1215 12:38:46.888585 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28\": container with ID starting with 6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28 not found: ID does not exist" containerID="6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.888611 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28"} err="failed to get container status \"6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28\": rpc error: code = NotFound desc = could not find container \"6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28\": container with ID starting with 6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28 not found: ID does not exist" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.888625 4719 scope.go:117] "RemoveContainer" containerID="2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.888811 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558"} err="failed to get container status \"2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558\": rpc error: code = NotFound desc = could not find container \"2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558\": container with ID starting with 2080949f016911f1e28b2090ac4da326f6c46ef067988e1b3dbd652793a23558 not found: ID does not exist" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.888831 4719 scope.go:117] "RemoveContainer" containerID="6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.889097 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28"} err="failed to get container status \"6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28\": rpc error: code = NotFound desc = could not find container \"6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28\": container with ID starting with 6010514a78221103f39fdb27c6d5646429dca785fc2331d064c0d16b25103e28 not found: ID does not exist" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.950635 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.960425 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.979249 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:38:46 crc kubenswrapper[4719]: E1215 12:38:46.979611 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edd905bf-6348-446d-abe8-fb632760bc1f" containerName="sg-core" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.979628 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="edd905bf-6348-446d-abe8-fb632760bc1f" containerName="sg-core" Dec 15 12:38:46 crc kubenswrapper[4719]: E1215 12:38:46.979643 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58fea5e6-9ff9-4a44-be8a-311735af3e4b" containerName="barbican-db-sync" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.979649 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="58fea5e6-9ff9-4a44-be8a-311735af3e4b" containerName="barbican-db-sync" Dec 15 12:38:46 crc kubenswrapper[4719]: E1215 12:38:46.979669 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edd905bf-6348-446d-abe8-fb632760bc1f" containerName="proxy-httpd" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.979674 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="edd905bf-6348-446d-abe8-fb632760bc1f" containerName="proxy-httpd" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.979848 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="edd905bf-6348-446d-abe8-fb632760bc1f" containerName="sg-core" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.979892 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="58fea5e6-9ff9-4a44-be8a-311735af3e4b" containerName="barbican-db-sync" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.979907 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="edd905bf-6348-446d-abe8-fb632760bc1f" containerName="proxy-httpd" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.981420 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.989169 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.989884 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 15 12:38:46 crc kubenswrapper[4719]: I1215 12:38:46.998786 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.069219 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-config-data\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.069624 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.069701 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-run-httpd\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.069755 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-scripts\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.069817 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnwhp\" (UniqueName: \"kubernetes.io/projected/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-kube-api-access-dnwhp\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.069843 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.069902 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-log-httpd\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.172634 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-scripts\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.172694 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnwhp\" (UniqueName: \"kubernetes.io/projected/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-kube-api-access-dnwhp\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.172714 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.172741 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-log-httpd\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.172776 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-config-data\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.172835 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.172904 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-run-httpd\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.173403 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-run-httpd\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.173658 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-log-httpd\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.183148 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-scripts\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.184257 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.184729 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.189991 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-config-data\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.221312 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnwhp\" (UniqueName: \"kubernetes.io/projected/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-kube-api-access-dnwhp\") pod \"ceilometer-0\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.233402 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6b87454b66-vfv5x"] Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.234920 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.238673 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.261076 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.267912 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-769f77d87-tfk84"] Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.269499 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.273134 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.273334 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-4dhqg" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.290945 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-769f77d87-tfk84"] Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.301295 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.308526 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6b87454b66-vfv5x"] Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.341458 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-48lqv"] Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.342784 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.377142 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwjrr\" (UniqueName: \"kubernetes.io/projected/d2f1a60c-ffe6-4563-9466-2a3129be55ee-kube-api-access-hwjrr\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.377212 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2f1a60c-ffe6-4563-9466-2a3129be55ee-logs\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.377252 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da95a03-4fe4-4674-b56c-f697d4ab3045-config-data\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.377300 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f1a60c-ffe6-4563-9466-2a3129be55ee-combined-ca-bundle\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.377333 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2da95a03-4fe4-4674-b56c-f697d4ab3045-logs\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.377371 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2f1a60c-ffe6-4563-9466-2a3129be55ee-config-data-custom\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.377431 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skhgl\" (UniqueName: \"kubernetes.io/projected/2da95a03-4fe4-4674-b56c-f697d4ab3045-kube-api-access-skhgl\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.377453 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f1a60c-ffe6-4563-9466-2a3129be55ee-config-data\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.377484 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da95a03-4fe4-4674-b56c-f697d4ab3045-combined-ca-bundle\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.377554 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2da95a03-4fe4-4674-b56c-f697d4ab3045-config-data-custom\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.389328 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.457009 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-48lqv"] Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.481520 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59cjb\" (UniqueName: \"kubernetes.io/projected/bae345c7-80a6-461b-88fd-51d1243eb2bd-kube-api-access-59cjb\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.481770 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.481802 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skhgl\" (UniqueName: \"kubernetes.io/projected/2da95a03-4fe4-4674-b56c-f697d4ab3045-kube-api-access-skhgl\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.481822 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f1a60c-ffe6-4563-9466-2a3129be55ee-config-data\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.481843 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-config\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.481910 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da95a03-4fe4-4674-b56c-f697d4ab3045-combined-ca-bundle\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.481930 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.481957 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.481995 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2da95a03-4fe4-4674-b56c-f697d4ab3045-config-data-custom\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.482014 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.482029 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwjrr\" (UniqueName: \"kubernetes.io/projected/d2f1a60c-ffe6-4563-9466-2a3129be55ee-kube-api-access-hwjrr\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.482049 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2f1a60c-ffe6-4563-9466-2a3129be55ee-logs\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.482076 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da95a03-4fe4-4674-b56c-f697d4ab3045-config-data\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.482099 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f1a60c-ffe6-4563-9466-2a3129be55ee-combined-ca-bundle\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.482130 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2da95a03-4fe4-4674-b56c-f697d4ab3045-logs\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.482145 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2f1a60c-ffe6-4563-9466-2a3129be55ee-config-data-custom\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.483688 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2f1a60c-ffe6-4563-9466-2a3129be55ee-logs\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.485142 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2da95a03-4fe4-4674-b56c-f697d4ab3045-logs\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.495395 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f1a60c-ffe6-4563-9466-2a3129be55ee-combined-ca-bundle\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.496205 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f1a60c-ffe6-4563-9466-2a3129be55ee-config-data\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.501082 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edd905bf-6348-446d-abe8-fb632760bc1f" path="/var/lib/kubelet/pods/edd905bf-6348-446d-abe8-fb632760bc1f/volumes" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.515631 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da95a03-4fe4-4674-b56c-f697d4ab3045-combined-ca-bundle\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.512409 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2f1a60c-ffe6-4563-9466-2a3129be55ee-config-data-custom\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.517034 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2da95a03-4fe4-4674-b56c-f697d4ab3045-config-data-custom\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.538269 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da95a03-4fe4-4674-b56c-f697d4ab3045-config-data\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.540253 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skhgl\" (UniqueName: \"kubernetes.io/projected/2da95a03-4fe4-4674-b56c-f697d4ab3045-kube-api-access-skhgl\") pod \"barbican-keystone-listener-6b87454b66-vfv5x\" (UID: \"2da95a03-4fe4-4674-b56c-f697d4ab3045\") " pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.551124 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwjrr\" (UniqueName: \"kubernetes.io/projected/d2f1a60c-ffe6-4563-9466-2a3129be55ee-kube-api-access-hwjrr\") pod \"barbican-worker-769f77d87-tfk84\" (UID: \"d2f1a60c-ffe6-4563-9466-2a3129be55ee\") " pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.576423 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc5fb48-jqf6b" podUID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.583781 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.583943 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59cjb\" (UniqueName: \"kubernetes.io/projected/bae345c7-80a6-461b-88fd-51d1243eb2bd-kube-api-access-59cjb\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.583977 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.584023 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-config\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.584048 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.584084 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.585025 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.585942 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.586441 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.588424 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-config\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.589356 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.636674 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59cjb\" (UniqueName: \"kubernetes.io/projected/bae345c7-80a6-461b-88fd-51d1243eb2bd-kube-api-access-59cjb\") pod \"dnsmasq-dns-59d5ff467f-48lqv\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.648354 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.727967 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c86c57db-47pxb"] Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.748186 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.793911 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.797545 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c86c57db-47pxb"] Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.799283 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-769f77d87-tfk84" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.883448 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.902122 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjtcn\" (UniqueName: \"kubernetes.io/projected/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-kube-api-access-rjtcn\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.902176 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data-custom\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.902245 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-combined-ca-bundle\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.902263 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:47 crc kubenswrapper[4719]: I1215 12:38:47.902319 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-logs\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.004813 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-combined-ca-bundle\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.004873 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.004940 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-logs\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.004975 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjtcn\" (UniqueName: \"kubernetes.io/projected/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-kube-api-access-rjtcn\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.005007 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data-custom\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.006895 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-logs\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.014299 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-combined-ca-bundle\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.026873 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.027635 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data-custom\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.038373 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjtcn\" (UniqueName: \"kubernetes.io/projected/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-kube-api-access-rjtcn\") pod \"barbican-api-6c86c57db-47pxb\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.307662 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.596998 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.610236 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.733926 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c89dt\" (UniqueName: \"kubernetes.io/projected/3243ab84-0a4c-4c74-83de-693c1c96e89e-kube-api-access-c89dt\") pod \"3243ab84-0a4c-4c74-83de-693c1c96e89e\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.734115 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-config\") pod \"3243ab84-0a4c-4c74-83de-693c1c96e89e\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.734161 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-combined-ca-bundle\") pod \"3243ab84-0a4c-4c74-83de-693c1c96e89e\" (UID: \"3243ab84-0a4c-4c74-83de-693c1c96e89e\") " Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.748146 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3243ab84-0a4c-4c74-83de-693c1c96e89e-kube-api-access-c89dt" (OuterVolumeSpecName: "kube-api-access-c89dt") pod "3243ab84-0a4c-4c74-83de-693c1c96e89e" (UID: "3243ab84-0a4c-4c74-83de-693c1c96e89e"). InnerVolumeSpecName "kube-api-access-c89dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.770970 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3243ab84-0a4c-4c74-83de-693c1c96e89e" (UID: "3243ab84-0a4c-4c74-83de-693c1c96e89e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.796941 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-config" (OuterVolumeSpecName: "config") pod "3243ab84-0a4c-4c74-83de-693c1c96e89e" (UID: "3243ab84-0a4c-4c74-83de-693c1c96e89e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.814762 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6b87454b66-vfv5x"] Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.835515 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.835803 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c89dt\" (UniqueName: \"kubernetes.io/projected/3243ab84-0a4c-4c74-83de-693c1c96e89e-kube-api-access-c89dt\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.835819 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3243ab84-0a4c-4c74-83de-693c1c96e89e-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.953706 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9vwxh" event={"ID":"3243ab84-0a4c-4c74-83de-693c1c96e89e","Type":"ContainerDied","Data":"0ef71242f60fe5967eab9ab9f67dffd198606eaa844feb54153501caa6272b83"} Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.953741 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ef71242f60fe5967eab9ab9f67dffd198606eaa844feb54153501caa6272b83" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.953795 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9vwxh" Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.958362 4719 generic.go:334] "Generic (PLEG): container finished" podID="ef9ed6e9-3050-4335-8ffc-97f889eb63e6" containerID="68928debcaf4ded9eda72c222dfdb8418b8bf08b43ca2b04dfd19ea947b50a35" exitCode=0 Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.958414 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gprbv" event={"ID":"ef9ed6e9-3050-4335-8ffc-97f889eb63e6","Type":"ContainerDied","Data":"68928debcaf4ded9eda72c222dfdb8418b8bf08b43ca2b04dfd19ea947b50a35"} Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.959588 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" event={"ID":"2da95a03-4fe4-4674-b56c-f697d4ab3045","Type":"ContainerStarted","Data":"cea0426c3173a794e181d61798a6a8c45304071a24de6a04647ea8e56c3bdd63"} Dec 15 12:38:48 crc kubenswrapper[4719]: I1215 12:38:48.973593 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf7e6ad7-6983-4455-af5d-5f7dda131dd2","Type":"ContainerStarted","Data":"e2c91729a53eae89759a916dd91f234e3df5233e9eb0be468093f37b8fc66fb0"} Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.027927 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-48lqv"] Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.047818 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-769f77d87-tfk84"] Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.146483 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-48lqv"] Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.165164 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-cblg6"] Dec 15 12:38:49 crc kubenswrapper[4719]: E1215 12:38:49.170180 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3243ab84-0a4c-4c74-83de-693c1c96e89e" containerName="neutron-db-sync" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.170208 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="3243ab84-0a4c-4c74-83de-693c1c96e89e" containerName="neutron-db-sync" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.170396 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="3243ab84-0a4c-4c74-83de-693c1c96e89e" containerName="neutron-db-sync" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.171467 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.218789 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-cblg6"] Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.249924 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.249981 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.274703 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.275124 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-config\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.275274 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.275349 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qdkv\" (UniqueName: \"kubernetes.io/projected/898ea52b-4011-43be-9f74-b07f9ae6cc38-kube-api-access-9qdkv\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.268057 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c86c57db-47pxb"] Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.376991 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.377065 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qdkv\" (UniqueName: \"kubernetes.io/projected/898ea52b-4011-43be-9f74-b07f9ae6cc38-kube-api-access-9qdkv\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.377122 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.377151 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.377186 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.377269 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-config\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.378266 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-config\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.378953 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.379597 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.380688 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.385238 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.410117 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qdkv\" (UniqueName: \"kubernetes.io/projected/898ea52b-4011-43be-9f74-b07f9ae6cc38-kube-api-access-9qdkv\") pod \"dnsmasq-dns-75c8ddd69c-cblg6\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.504164 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.544470 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5649967964-ctb22"] Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.545730 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.559234 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.559479 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.568642 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-g5z85" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.603333 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.677140 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5649967964-ctb22"] Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.710346 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-config\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.710400 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-combined-ca-bundle\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.710440 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtrjk\" (UniqueName: \"kubernetes.io/projected/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-kube-api-access-wtrjk\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.710494 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-httpd-config\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.710563 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-ovndb-tls-certs\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.813734 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-ovndb-tls-certs\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.813811 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-config\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.813839 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-combined-ca-bundle\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.813919 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtrjk\" (UniqueName: \"kubernetes.io/projected/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-kube-api-access-wtrjk\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.814003 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-httpd-config\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.820127 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-config\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.820318 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-httpd-config\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.821260 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-ovndb-tls-certs\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.834795 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-combined-ca-bundle\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.854539 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtrjk\" (UniqueName: \"kubernetes.io/projected/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-kube-api-access-wtrjk\") pod \"neutron-5649967964-ctb22\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:49 crc kubenswrapper[4719]: I1215 12:38:49.916667 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.025211 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-769f77d87-tfk84" event={"ID":"d2f1a60c-ffe6-4563-9466-2a3129be55ee","Type":"ContainerStarted","Data":"7f4245c56e4ff1310e22d9ae211984e9d1cd0801a784eb961ed0c575eb501415"} Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.041494 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86c57db-47pxb" event={"ID":"7818f1dd-1db1-4af7-b833-e4bbc91e9a09","Type":"ContainerStarted","Data":"ebff589ae4bfeda370095f36ab298c3c5529a0314fa5beb1b538094a10e05edf"} Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.041546 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86c57db-47pxb" event={"ID":"7818f1dd-1db1-4af7-b833-e4bbc91e9a09","Type":"ContainerStarted","Data":"ff02c273660cef618806136d485846099b778e88c40235e63f0a6b6cc291b13d"} Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.044477 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf7e6ad7-6983-4455-af5d-5f7dda131dd2","Type":"ContainerStarted","Data":"d67ccc630a371b702f7cba088063782b8442732e4ad34971ee7482ddb314d390"} Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.047796 4719 generic.go:334] "Generic (PLEG): container finished" podID="bae345c7-80a6-461b-88fd-51d1243eb2bd" containerID="b9cfca9de9b8431b6ce934998327708959eef2fbadf41d33b84d26f43ffa6d8c" exitCode=0 Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.047997 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" event={"ID":"bae345c7-80a6-461b-88fd-51d1243eb2bd","Type":"ContainerDied","Data":"b9cfca9de9b8431b6ce934998327708959eef2fbadf41d33b84d26f43ffa6d8c"} Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.048022 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" event={"ID":"bae345c7-80a6-461b-88fd-51d1243eb2bd","Type":"ContainerStarted","Data":"b32863636346b972141d5f90ee15d1429ac31d1f987d4ff426d92b378bdd1a62"} Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.274590 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-cblg6"] Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.903106 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.957537 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-sb\") pod \"bae345c7-80a6-461b-88fd-51d1243eb2bd\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.957607 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59cjb\" (UniqueName: \"kubernetes.io/projected/bae345c7-80a6-461b-88fd-51d1243eb2bd-kube-api-access-59cjb\") pod \"bae345c7-80a6-461b-88fd-51d1243eb2bd\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.957681 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-config\") pod \"bae345c7-80a6-461b-88fd-51d1243eb2bd\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.957777 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-swift-storage-0\") pod \"bae345c7-80a6-461b-88fd-51d1243eb2bd\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.957843 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-svc\") pod \"bae345c7-80a6-461b-88fd-51d1243eb2bd\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.957875 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-nb\") pod \"bae345c7-80a6-461b-88fd-51d1243eb2bd\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " Dec 15 12:38:50 crc kubenswrapper[4719]: I1215 12:38:50.998325 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae345c7-80a6-461b-88fd-51d1243eb2bd-kube-api-access-59cjb" (OuterVolumeSpecName: "kube-api-access-59cjb") pod "bae345c7-80a6-461b-88fd-51d1243eb2bd" (UID: "bae345c7-80a6-461b-88fd-51d1243eb2bd"). InnerVolumeSpecName "kube-api-access-59cjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.057662 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5649967964-ctb22"] Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.070775 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-config" (OuterVolumeSpecName: "config") pod "bae345c7-80a6-461b-88fd-51d1243eb2bd" (UID: "bae345c7-80a6-461b-88fd-51d1243eb2bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.071239 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bae345c7-80a6-461b-88fd-51d1243eb2bd" (UID: "bae345c7-80a6-461b-88fd-51d1243eb2bd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.071544 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-nb\") pod \"bae345c7-80a6-461b-88fd-51d1243eb2bd\" (UID: \"bae345c7-80a6-461b-88fd-51d1243eb2bd\") " Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.071777 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bae345c7-80a6-461b-88fd-51d1243eb2bd" (UID: "bae345c7-80a6-461b-88fd-51d1243eb2bd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: W1215 12:38:51.071929 4719 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/bae345c7-80a6-461b-88fd-51d1243eb2bd/volumes/kubernetes.io~configmap/ovsdbserver-nb Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.071938 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bae345c7-80a6-461b-88fd-51d1243eb2bd" (UID: "bae345c7-80a6-461b-88fd-51d1243eb2bd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.072022 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59cjb\" (UniqueName: \"kubernetes.io/projected/bae345c7-80a6-461b-88fd-51d1243eb2bd-kube-api-access-59cjb\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.072035 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.072045 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.072053 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.096074 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bae345c7-80a6-461b-88fd-51d1243eb2bd" (UID: "bae345c7-80a6-461b-88fd-51d1243eb2bd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.099385 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bae345c7-80a6-461b-88fd-51d1243eb2bd" (UID: "bae345c7-80a6-461b-88fd-51d1243eb2bd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.100908 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gprbv" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.101316 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" event={"ID":"898ea52b-4011-43be-9f74-b07f9ae6cc38","Type":"ContainerStarted","Data":"ffa7725f4818c2e8850447381847d5d5f1f68f86ce1869214b19098c5e016082"} Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.113433 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" event={"ID":"bae345c7-80a6-461b-88fd-51d1243eb2bd","Type":"ContainerDied","Data":"b32863636346b972141d5f90ee15d1429ac31d1f987d4ff426d92b378bdd1a62"} Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.113479 4719 scope.go:117] "RemoveContainer" containerID="b9cfca9de9b8431b6ce934998327708959eef2fbadf41d33b84d26f43ffa6d8c" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.113609 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-48lqv" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.153431 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gprbv" event={"ID":"ef9ed6e9-3050-4335-8ffc-97f889eb63e6","Type":"ContainerDied","Data":"d4ab5662a225660f3af235a30327f57713cd8d78e9122e59a3eba3233b4ddf73"} Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.153469 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4ab5662a225660f3af235a30327f57713cd8d78e9122e59a3eba3233b4ddf73" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.153530 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gprbv" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.158239 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86c57db-47pxb" event={"ID":"7818f1dd-1db1-4af7-b833-e4bbc91e9a09","Type":"ContainerStarted","Data":"28782885e6f9c77a44600262069f881d5a034ddbee91fa23088db779f8574dfe"} Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.159178 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.159201 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.175807 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-etc-machine-id\") pod \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.175910 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-scripts\") pod \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.175937 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-db-sync-config-data\") pod \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.175976 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-config-data\") pod \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.176048 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-combined-ca-bundle\") pod \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.176069 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4drnb\" (UniqueName: \"kubernetes.io/projected/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-kube-api-access-4drnb\") pod \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\" (UID: \"ef9ed6e9-3050-4335-8ffc-97f889eb63e6\") " Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.176457 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.176471 4719 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bae345c7-80a6-461b-88fd-51d1243eb2bd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.179689 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ef9ed6e9-3050-4335-8ffc-97f889eb63e6" (UID: "ef9ed6e9-3050-4335-8ffc-97f889eb63e6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.182087 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-kube-api-access-4drnb" (OuterVolumeSpecName: "kube-api-access-4drnb") pod "ef9ed6e9-3050-4335-8ffc-97f889eb63e6" (UID: "ef9ed6e9-3050-4335-8ffc-97f889eb63e6"). InnerVolumeSpecName "kube-api-access-4drnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.191648 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ef9ed6e9-3050-4335-8ffc-97f889eb63e6" (UID: "ef9ed6e9-3050-4335-8ffc-97f889eb63e6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.205977 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-scripts" (OuterVolumeSpecName: "scripts") pod "ef9ed6e9-3050-4335-8ffc-97f889eb63e6" (UID: "ef9ed6e9-3050-4335-8ffc-97f889eb63e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.244741 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-48lqv"] Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.264136 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef9ed6e9-3050-4335-8ffc-97f889eb63e6" (UID: "ef9ed6e9-3050-4335-8ffc-97f889eb63e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.287007 4719 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.287036 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.287045 4719 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.287055 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.287063 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4drnb\" (UniqueName: \"kubernetes.io/projected/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-kube-api-access-4drnb\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.289279 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-48lqv"] Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.304905 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c86c57db-47pxb" podStartSLOduration=4.304880199 podStartE2EDuration="4.304880199s" podCreationTimestamp="2025-12-15 12:38:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:38:51.264582226 +0000 UTC m=+1292.206875256" watchObservedRunningTime="2025-12-15 12:38:51.304880199 +0000 UTC m=+1292.247173229" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.344057 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-config-data" (OuterVolumeSpecName: "config-data") pod "ef9ed6e9-3050-4335-8ffc-97f889eb63e6" (UID: "ef9ed6e9-3050-4335-8ffc-97f889eb63e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.357292 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.357342 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.390885 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef9ed6e9-3050-4335-8ffc-97f889eb63e6-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:51 crc kubenswrapper[4719]: I1215 12:38:51.490914 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bae345c7-80a6-461b-88fd-51d1243eb2bd" path="/var/lib/kubelet/pods/bae345c7-80a6-461b-88fd-51d1243eb2bd/volumes" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.255134 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5649967964-ctb22" event={"ID":"2a1fc07a-9062-4ed7-8610-065fa84f1b2b","Type":"ContainerStarted","Data":"594ed6de5e5587a3240d090ec7dcad0447ce09468e5b92aef60bbc735df86854"} Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.255509 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5649967964-ctb22" event={"ID":"2a1fc07a-9062-4ed7-8610-065fa84f1b2b","Type":"ContainerStarted","Data":"6c8f61185a298bc47415e2490728b2af74efe149d868a8e1c30bbe335d4b351b"} Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.255530 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5649967964-ctb22" event={"ID":"2a1fc07a-9062-4ed7-8610-065fa84f1b2b","Type":"ContainerStarted","Data":"b7f4cef3a00f011ddf87b7a272684ad089ce3cba6f665e4deaf96c24edb74ae9"} Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.255593 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5649967964-ctb22" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.261816 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf7e6ad7-6983-4455-af5d-5f7dda131dd2","Type":"ContainerStarted","Data":"3cb74a676d33c8c35242627d7a2d15a9dab16f3673e70b096107042b68e7395c"} Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.263549 4719 generic.go:334] "Generic (PLEG): container finished" podID="898ea52b-4011-43be-9f74-b07f9ae6cc38" containerID="b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765" exitCode=0 Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.264477 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" event={"ID":"898ea52b-4011-43be-9f74-b07f9ae6cc38","Type":"ContainerDied","Data":"b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765"} Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.287349 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5649967964-ctb22" podStartSLOduration=3.28732856 podStartE2EDuration="3.28732856s" podCreationTimestamp="2025-12-15 12:38:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:38:52.282211361 +0000 UTC m=+1293.224504411" watchObservedRunningTime="2025-12-15 12:38:52.28732856 +0000 UTC m=+1293.229621600" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.455717 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 15 12:38:52 crc kubenswrapper[4719]: E1215 12:38:52.456111 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9ed6e9-3050-4335-8ffc-97f889eb63e6" containerName="cinder-db-sync" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.456130 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9ed6e9-3050-4335-8ffc-97f889eb63e6" containerName="cinder-db-sync" Dec 15 12:38:52 crc kubenswrapper[4719]: E1215 12:38:52.456143 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae345c7-80a6-461b-88fd-51d1243eb2bd" containerName="init" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.456149 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae345c7-80a6-461b-88fd-51d1243eb2bd" containerName="init" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.456310 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef9ed6e9-3050-4335-8ffc-97f889eb63e6" containerName="cinder-db-sync" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.456332 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae345c7-80a6-461b-88fd-51d1243eb2bd" containerName="init" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.457194 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.460424 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dwxb2" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.460560 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.460808 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.462587 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.482910 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.525940 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.526302 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5909a62d-a8e1-4a30-bb53-4d68e403f090-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.526448 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.526672 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-scripts\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.526883 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.526953 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfmvp\" (UniqueName: \"kubernetes.io/projected/5909a62d-a8e1-4a30-bb53-4d68e403f090-kube-api-access-dfmvp\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.559974 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-cblg6"] Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.629178 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.629226 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5909a62d-a8e1-4a30-bb53-4d68e403f090-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.629260 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.629326 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-scripts\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.629345 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.629372 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfmvp\" (UniqueName: \"kubernetes.io/projected/5909a62d-a8e1-4a30-bb53-4d68e403f090-kube-api-access-dfmvp\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.630519 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5909a62d-a8e1-4a30-bb53-4d68e403f090-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.635623 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-scripts\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.636544 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.650887 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.651395 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfmvp\" (UniqueName: \"kubernetes.io/projected/5909a62d-a8e1-4a30-bb53-4d68e403f090-kube-api-access-dfmvp\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.656666 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-b4mfs"] Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.657831 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data\") pod \"cinder-scheduler-0\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.657995 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.685941 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-b4mfs"] Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.730333 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-svc\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.730373 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.730393 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-config\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.730407 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvldx\" (UniqueName: \"kubernetes.io/projected/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-kube-api-access-qvldx\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.730433 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.730465 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.782611 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.818766 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.820276 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.825058 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.832181 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-svc\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.832228 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.832249 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-config\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.832264 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvldx\" (UniqueName: \"kubernetes.io/projected/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-kube-api-access-qvldx\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.832294 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.832327 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.833441 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.833584 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-svc\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.834426 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-config\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.835033 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.837820 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.840536 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.876674 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvldx\" (UniqueName: \"kubernetes.io/projected/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-kube-api-access-qvldx\") pod \"dnsmasq-dns-5784cf869f-b4mfs\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.934279 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.934316 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adf32d03-4559-4d49-8577-c0dac00c1fec-logs\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.934540 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-scripts\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.934605 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftjtb\" (UniqueName: \"kubernetes.io/projected/adf32d03-4559-4d49-8577-c0dac00c1fec-kube-api-access-ftjtb\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.934623 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.934686 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adf32d03-4559-4d49-8577-c0dac00c1fec-etc-machine-id\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:52 crc kubenswrapper[4719]: I1215 12:38:52.934730 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data-custom\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.030363 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.036143 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adf32d03-4559-4d49-8577-c0dac00c1fec-etc-machine-id\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.036262 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adf32d03-4559-4d49-8577-c0dac00c1fec-etc-machine-id\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.036556 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data-custom\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.036772 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.036813 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adf32d03-4559-4d49-8577-c0dac00c1fec-logs\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.036891 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-scripts\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.036916 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.036933 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftjtb\" (UniqueName: \"kubernetes.io/projected/adf32d03-4559-4d49-8577-c0dac00c1fec-kube-api-access-ftjtb\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.037102 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adf32d03-4559-4d49-8577-c0dac00c1fec-logs\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.041089 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.041997 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data-custom\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.042593 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-scripts\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.053914 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.054410 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftjtb\" (UniqueName: \"kubernetes.io/projected/adf32d03-4559-4d49-8577-c0dac00c1fec-kube-api-access-ftjtb\") pod \"cinder-api-0\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " pod="openstack/cinder-api-0" Dec 15 12:38:53 crc kubenswrapper[4719]: I1215 12:38:53.135559 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 15 12:38:54 crc kubenswrapper[4719]: I1215 12:38:54.320393 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf7e6ad7-6983-4455-af5d-5f7dda131dd2","Type":"ContainerStarted","Data":"87aabd4d0941940cd26018e043a9abce5ca108f30d47e257c61cb6570c697f40"} Dec 15 12:38:55 crc kubenswrapper[4719]: I1215 12:38:55.262405 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 15 12:38:55 crc kubenswrapper[4719]: I1215 12:38:55.375695 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adf32d03-4559-4d49-8577-c0dac00c1fec","Type":"ContainerStarted","Data":"8119c821b57b0b70ed1a77e598c860afdda1bc42a29d05231cdc2739ae30247a"} Dec 15 12:38:55 crc kubenswrapper[4719]: I1215 12:38:55.387545 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" event={"ID":"2da95a03-4fe4-4674-b56c-f697d4ab3045","Type":"ContainerStarted","Data":"beb46cc44e22d575940e2dd5086ac24a70d2a6d713afc35a5643ad4854280654"} Dec 15 12:38:55 crc kubenswrapper[4719]: I1215 12:38:55.457239 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" event={"ID":"898ea52b-4011-43be-9f74-b07f9ae6cc38","Type":"ContainerStarted","Data":"0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b"} Dec 15 12:38:55 crc kubenswrapper[4719]: I1215 12:38:55.457360 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" podUID="898ea52b-4011-43be-9f74-b07f9ae6cc38" containerName="dnsmasq-dns" containerID="cri-o://0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b" gracePeriod=10 Dec 15 12:38:55 crc kubenswrapper[4719]: I1215 12:38:55.457548 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:55 crc kubenswrapper[4719]: I1215 12:38:55.513175 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" podStartSLOduration=6.513157376 podStartE2EDuration="6.513157376s" podCreationTimestamp="2025-12-15 12:38:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:38:55.492676204 +0000 UTC m=+1296.434969234" watchObservedRunningTime="2025-12-15 12:38:55.513157376 +0000 UTC m=+1296.455450406" Dec 15 12:38:55 crc kubenswrapper[4719]: W1215 12:38:55.636000 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5909a62d_a8e1_4a30_bb53_4d68e403f090.slice/crio-98df4ad8100837012f11e69a3e57809575174ce33f9c982906943e643d6eb1ec WatchSource:0}: Error finding container 98df4ad8100837012f11e69a3e57809575174ce33f9c982906943e643d6eb1ec: Status 404 returned error can't find the container with id 98df4ad8100837012f11e69a3e57809575174ce33f9c982906943e643d6eb1ec Dec 15 12:38:55 crc kubenswrapper[4719]: W1215 12:38:55.637090 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd68b9a30_55fe_4d32_8cba_9a12dcdc210c.slice/crio-ba157ab11cfe480035a4cf889637d80cb7113976b48c930cf382a7661c7c0852 WatchSource:0}: Error finding container ba157ab11cfe480035a4cf889637d80cb7113976b48c930cf382a7661c7c0852: Status 404 returned error can't find the container with id ba157ab11cfe480035a4cf889637d80cb7113976b48c930cf382a7661c7c0852 Dec 15 12:38:55 crc kubenswrapper[4719]: I1215 12:38:55.641187 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-769f77d87-tfk84" event={"ID":"d2f1a60c-ffe6-4563-9466-2a3129be55ee","Type":"ContainerStarted","Data":"f85184be35cb232ebed78c1a2041d40656fe12a5ea7386c23f429665c42e6bf1"} Dec 15 12:38:55 crc kubenswrapper[4719]: I1215 12:38:55.641233 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 15 12:38:55 crc kubenswrapper[4719]: I1215 12:38:55.641254 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-b4mfs"] Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.278598 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.305552 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.411026 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-sb\") pod \"898ea52b-4011-43be-9f74-b07f9ae6cc38\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.411345 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-swift-storage-0\") pod \"898ea52b-4011-43be-9f74-b07f9ae6cc38\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.411384 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-nb\") pod \"898ea52b-4011-43be-9f74-b07f9ae6cc38\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.411461 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-config\") pod \"898ea52b-4011-43be-9f74-b07f9ae6cc38\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.411513 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-svc\") pod \"898ea52b-4011-43be-9f74-b07f9ae6cc38\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.411610 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qdkv\" (UniqueName: \"kubernetes.io/projected/898ea52b-4011-43be-9f74-b07f9ae6cc38-kube-api-access-9qdkv\") pod \"898ea52b-4011-43be-9f74-b07f9ae6cc38\" (UID: \"898ea52b-4011-43be-9f74-b07f9ae6cc38\") " Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.479030 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/898ea52b-4011-43be-9f74-b07f9ae6cc38-kube-api-access-9qdkv" (OuterVolumeSpecName: "kube-api-access-9qdkv") pod "898ea52b-4011-43be-9f74-b07f9ae6cc38" (UID: "898ea52b-4011-43be-9f74-b07f9ae6cc38"). InnerVolumeSpecName "kube-api-access-9qdkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.514033 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qdkv\" (UniqueName: \"kubernetes.io/projected/898ea52b-4011-43be-9f74-b07f9ae6cc38-kube-api-access-9qdkv\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.526878 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" event={"ID":"2da95a03-4fe4-4674-b56c-f697d4ab3045","Type":"ContainerStarted","Data":"d858e2db2f0e4b2c9df40708c6d8c8d62067a8105c32f360757077f708cfc56c"} Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.529751 4719 generic.go:334] "Generic (PLEG): container finished" podID="898ea52b-4011-43be-9f74-b07f9ae6cc38" containerID="0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b" exitCode=0 Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.529798 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" event={"ID":"898ea52b-4011-43be-9f74-b07f9ae6cc38","Type":"ContainerDied","Data":"0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b"} Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.529818 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" event={"ID":"898ea52b-4011-43be-9f74-b07f9ae6cc38","Type":"ContainerDied","Data":"ffa7725f4818c2e8850447381847d5d5f1f68f86ce1869214b19098c5e016082"} Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.529833 4719 scope.go:117] "RemoveContainer" containerID="0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.529943 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-cblg6" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.536612 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" event={"ID":"d68b9a30-55fe-4d32-8cba-9a12dcdc210c","Type":"ContainerStarted","Data":"ba157ab11cfe480035a4cf889637d80cb7113976b48c930cf382a7661c7c0852"} Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.541401 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5909a62d-a8e1-4a30-bb53-4d68e403f090","Type":"ContainerStarted","Data":"98df4ad8100837012f11e69a3e57809575174ce33f9c982906943e643d6eb1ec"} Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.550903 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6b87454b66-vfv5x" podStartSLOduration=3.675009166 podStartE2EDuration="9.550889976s" podCreationTimestamp="2025-12-15 12:38:47 +0000 UTC" firstStartedPulling="2025-12-15 12:38:48.817268962 +0000 UTC m=+1289.759561992" lastFinishedPulling="2025-12-15 12:38:54.693149772 +0000 UTC m=+1295.635442802" observedRunningTime="2025-12-15 12:38:56.543967879 +0000 UTC m=+1297.486260909" watchObservedRunningTime="2025-12-15 12:38:56.550889976 +0000 UTC m=+1297.493182996" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.590040 4719 scope.go:117] "RemoveContainer" containerID="b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.682703 4719 scope.go:117] "RemoveContainer" containerID="0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b" Dec 15 12:38:56 crc kubenswrapper[4719]: E1215 12:38:56.689004 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b\": container with ID starting with 0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b not found: ID does not exist" containerID="0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.689056 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b"} err="failed to get container status \"0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b\": rpc error: code = NotFound desc = could not find container \"0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b\": container with ID starting with 0c6062f925dcc5f1ffa5a4041b4083212ad10fdb1d38460c44af63aafde9f53b not found: ID does not exist" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.689082 4719 scope.go:117] "RemoveContainer" containerID="b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765" Dec 15 12:38:56 crc kubenswrapper[4719]: E1215 12:38:56.691232 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765\": container with ID starting with b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765 not found: ID does not exist" containerID="b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.691258 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765"} err="failed to get container status \"b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765\": rpc error: code = NotFound desc = could not find container \"b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765\": container with ID starting with b5f7c489bbf9d1322b7d382b6fbe3b5a62ad8a93251b6530fcb514c8833e9765 not found: ID does not exist" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.976630 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "898ea52b-4011-43be-9f74-b07f9ae6cc38" (UID: "898ea52b-4011-43be-9f74-b07f9ae6cc38"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.982980 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-config" (OuterVolumeSpecName: "config") pod "898ea52b-4011-43be-9f74-b07f9ae6cc38" (UID: "898ea52b-4011-43be-9f74-b07f9ae6cc38"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.989461 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "898ea52b-4011-43be-9f74-b07f9ae6cc38" (UID: "898ea52b-4011-43be-9f74-b07f9ae6cc38"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.994564 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "898ea52b-4011-43be-9f74-b07f9ae6cc38" (UID: "898ea52b-4011-43be-9f74-b07f9ae6cc38"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:56 crc kubenswrapper[4719]: I1215 12:38:56.995994 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "898ea52b-4011-43be-9f74-b07f9ae6cc38" (UID: "898ea52b-4011-43be-9f74-b07f9ae6cc38"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.030314 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.030356 4719 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.030369 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.030381 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.030392 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/898ea52b-4011-43be-9f74-b07f9ae6cc38-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.268931 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-cblg6"] Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.275327 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-cblg6"] Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.499385 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="898ea52b-4011-43be-9f74-b07f9ae6cc38" path="/var/lib/kubelet/pods/898ea52b-4011-43be-9f74-b07f9ae6cc38/volumes" Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.615462 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf7e6ad7-6983-4455-af5d-5f7dda131dd2","Type":"ContainerStarted","Data":"a3a301cb98da46f13ca27360ca1c76299fafc4f89aa2582e83f52a2477149de9"} Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.615689 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.640038 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-769f77d87-tfk84" event={"ID":"d2f1a60c-ffe6-4563-9466-2a3129be55ee","Type":"ContainerStarted","Data":"8bf8a89e21bf9d38b1c603cbf09e97ae2e05b83bf801b177772c373226ed3394"} Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.677073 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.149722019 podStartE2EDuration="11.677053227s" podCreationTimestamp="2025-12-15 12:38:46 +0000 UTC" firstStartedPulling="2025-12-15 12:38:48.616595531 +0000 UTC m=+1289.558888561" lastFinishedPulling="2025-12-15 12:38:56.143926739 +0000 UTC m=+1297.086219769" observedRunningTime="2025-12-15 12:38:57.66977552 +0000 UTC m=+1298.612068550" watchObservedRunningTime="2025-12-15 12:38:57.677053227 +0000 UTC m=+1298.619346257" Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.699072 4719 generic.go:334] "Generic (PLEG): container finished" podID="d68b9a30-55fe-4d32-8cba-9a12dcdc210c" containerID="e0d877200e56dde9fde96c37bd8c6403c6359b013c28cd758f916771d3a39598" exitCode=0 Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.699157 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" event={"ID":"d68b9a30-55fe-4d32-8cba-9a12dcdc210c","Type":"ContainerDied","Data":"e0d877200e56dde9fde96c37bd8c6403c6359b013c28cd758f916771d3a39598"} Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.736698 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adf32d03-4559-4d49-8577-c0dac00c1fec","Type":"ContainerStarted","Data":"45b06a2e47710e3a4c485c5df45b5a490e7e8faa8b68cc0e55dfd66411a9f377"} Dec 15 12:38:57 crc kubenswrapper[4719]: I1215 12:38:57.767949 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-769f77d87-tfk84" podStartSLOduration=5.121211269 podStartE2EDuration="10.767930347s" podCreationTimestamp="2025-12-15 12:38:47 +0000 UTC" firstStartedPulling="2025-12-15 12:38:49.020997987 +0000 UTC m=+1289.963291017" lastFinishedPulling="2025-12-15 12:38:54.667717065 +0000 UTC m=+1295.610010095" observedRunningTime="2025-12-15 12:38:57.735934873 +0000 UTC m=+1298.678227923" watchObservedRunningTime="2025-12-15 12:38:57.767930347 +0000 UTC m=+1298.710223377" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.157453 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5668b96c97-48pkv"] Dec 15 12:38:58 crc kubenswrapper[4719]: E1215 12:38:58.163182 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="898ea52b-4011-43be-9f74-b07f9ae6cc38" containerName="init" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.163215 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="898ea52b-4011-43be-9f74-b07f9ae6cc38" containerName="init" Dec 15 12:38:58 crc kubenswrapper[4719]: E1215 12:38:58.163238 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="898ea52b-4011-43be-9f74-b07f9ae6cc38" containerName="dnsmasq-dns" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.163244 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="898ea52b-4011-43be-9f74-b07f9ae6cc38" containerName="dnsmasq-dns" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.163519 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="898ea52b-4011-43be-9f74-b07f9ae6cc38" containerName="dnsmasq-dns" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.164429 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.172059 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.172387 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.191940 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5668b96c97-48pkv"] Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.287683 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-combined-ca-bundle\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.288061 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-public-tls-certs\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.288108 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-ovndb-tls-certs\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.288130 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-config\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.288187 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-httpd-config\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.288248 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-internal-tls-certs\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.288270 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5lqx\" (UniqueName: \"kubernetes.io/projected/4ddfe7f4-2192-4520-bfae-028a2c41e459-kube-api-access-d5lqx\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.390460 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-combined-ca-bundle\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.390594 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-public-tls-certs\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.390624 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-ovndb-tls-certs\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.390648 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-config\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.390720 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-httpd-config\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.390792 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-internal-tls-certs\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.390839 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5lqx\" (UniqueName: \"kubernetes.io/projected/4ddfe7f4-2192-4520-bfae-028a2c41e459-kube-api-access-d5lqx\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.395135 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-ovndb-tls-certs\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.399291 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-public-tls-certs\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.402736 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-config\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.403073 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-internal-tls-certs\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.418535 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-combined-ca-bundle\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.421444 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4ddfe7f4-2192-4520-bfae-028a2c41e459-httpd-config\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.427432 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5lqx\" (UniqueName: \"kubernetes.io/projected/4ddfe7f4-2192-4520-bfae-028a2c41e459-kube-api-access-d5lqx\") pod \"neutron-5668b96c97-48pkv\" (UID: \"4ddfe7f4-2192-4520-bfae-028a2c41e459\") " pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.572269 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.770234 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5909a62d-a8e1-4a30-bb53-4d68e403f090","Type":"ContainerStarted","Data":"da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b"} Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.805199 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" event={"ID":"d68b9a30-55fe-4d32-8cba-9a12dcdc210c","Type":"ContainerStarted","Data":"ff799566508d1613e80c3cd0de443e3cb223b7e9d5b48c1b857f9511ebaa7d30"} Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.806351 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.832304 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adf32d03-4559-4d49-8577-c0dac00c1fec","Type":"ContainerStarted","Data":"03a72fcca616f5e0c4506044788aeee363458eaccd166093db389ab297da0d56"} Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.832830 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerName="cinder-api-log" containerID="cri-o://45b06a2e47710e3a4c485c5df45b5a490e7e8faa8b68cc0e55dfd66411a9f377" gracePeriod=30 Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.832965 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.833282 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerName="cinder-api" containerID="cri-o://03a72fcca616f5e0c4506044788aeee363458eaccd166093db389ab297da0d56" gracePeriod=30 Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.843659 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" podStartSLOduration=6.843637676 podStartE2EDuration="6.843637676s" podCreationTimestamp="2025-12-15 12:38:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:38:58.832570179 +0000 UTC m=+1299.774863209" watchObservedRunningTime="2025-12-15 12:38:58.843637676 +0000 UTC m=+1299.785930706" Dec 15 12:38:58 crc kubenswrapper[4719]: I1215 12:38:58.866422 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.866386879 podStartE2EDuration="6.866386879s" podCreationTimestamp="2025-12-15 12:38:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:38:58.860798424 +0000 UTC m=+1299.803091454" watchObservedRunningTime="2025-12-15 12:38:58.866386879 +0000 UTC m=+1299.808679909" Dec 15 12:38:59 crc kubenswrapper[4719]: I1215 12:38:59.401513 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5668b96c97-48pkv"] Dec 15 12:38:59 crc kubenswrapper[4719]: I1215 12:38:59.854764 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5668b96c97-48pkv" event={"ID":"4ddfe7f4-2192-4520-bfae-028a2c41e459","Type":"ContainerStarted","Data":"41266954d5d0fcd7a5bf2f639f07b23b8fd000140c28ac386e349ddaef7b4433"} Dec 15 12:38:59 crc kubenswrapper[4719]: I1215 12:38:59.871685 4719 generic.go:334] "Generic (PLEG): container finished" podID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerID="45b06a2e47710e3a4c485c5df45b5a490e7e8faa8b68cc0e55dfd66411a9f377" exitCode=143 Dec 15 12:38:59 crc kubenswrapper[4719]: I1215 12:38:59.872465 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adf32d03-4559-4d49-8577-c0dac00c1fec","Type":"ContainerDied","Data":"45b06a2e47710e3a4c485c5df45b5a490e7e8faa8b68cc0e55dfd66411a9f377"} Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.060629 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7ff7556d66-bpsdl"] Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.062132 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.070989 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.071395 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.078029 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7ff7556d66-bpsdl"] Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.156819 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-internal-tls-certs\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.156938 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7zjw\" (UniqueName: \"kubernetes.io/projected/a996feea-1c4e-496f-a91b-2f07b852d22e-kube-api-access-w7zjw\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.156961 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-combined-ca-bundle\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.156998 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a996feea-1c4e-496f-a91b-2f07b852d22e-logs\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.157075 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-config-data\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.157114 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-config-data-custom\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.157139 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-public-tls-certs\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.259201 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-internal-tls-certs\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.260336 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7zjw\" (UniqueName: \"kubernetes.io/projected/a996feea-1c4e-496f-a91b-2f07b852d22e-kube-api-access-w7zjw\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.260358 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-combined-ca-bundle\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.260396 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a996feea-1c4e-496f-a91b-2f07b852d22e-logs\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.260657 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-config-data\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.261097 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-config-data-custom\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.261119 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-public-tls-certs\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.260976 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a996feea-1c4e-496f-a91b-2f07b852d22e-logs\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.264808 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-internal-tls-certs\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.264830 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-config-data-custom\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.265249 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-public-tls-certs\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.268516 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-config-data\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.269442 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a996feea-1c4e-496f-a91b-2f07b852d22e-combined-ca-bundle\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.287624 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7zjw\" (UniqueName: \"kubernetes.io/projected/a996feea-1c4e-496f-a91b-2f07b852d22e-kube-api-access-w7zjw\") pod \"barbican-api-7ff7556d66-bpsdl\" (UID: \"a996feea-1c4e-496f-a91b-2f07b852d22e\") " pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.432848 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.882305 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5668b96c97-48pkv" event={"ID":"4ddfe7f4-2192-4520-bfae-028a2c41e459","Type":"ContainerStarted","Data":"17b4aa4779bc9ff4006f1000c2b4a200aed20f272a20c1ff74991bab77400631"} Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.882533 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5668b96c97-48pkv" event={"ID":"4ddfe7f4-2192-4520-bfae-028a2c41e459","Type":"ContainerStarted","Data":"201a626f5baf8a04846ae7c283a271177607b11a835a048f3c6abe5d92258c23"} Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.883396 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.884757 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5909a62d-a8e1-4a30-bb53-4d68e403f090","Type":"ContainerStarted","Data":"a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142"} Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.906143 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5668b96c97-48pkv" podStartSLOduration=2.906127388 podStartE2EDuration="2.906127388s" podCreationTimestamp="2025-12-15 12:38:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:39:00.899649755 +0000 UTC m=+1301.841942785" watchObservedRunningTime="2025-12-15 12:39:00.906127388 +0000 UTC m=+1301.848420418" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.937132 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.348880374 podStartE2EDuration="8.9371153s" podCreationTimestamp="2025-12-15 12:38:52 +0000 UTC" firstStartedPulling="2025-12-15 12:38:55.642203132 +0000 UTC m=+1296.584496162" lastFinishedPulling="2025-12-15 12:38:57.230438058 +0000 UTC m=+1298.172731088" observedRunningTime="2025-12-15 12:39:00.927100435 +0000 UTC m=+1301.869393465" watchObservedRunningTime="2025-12-15 12:39:00.9371153 +0000 UTC m=+1301.879408330" Dec 15 12:39:00 crc kubenswrapper[4719]: I1215 12:39:00.998931 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7ff7556d66-bpsdl"] Dec 15 12:39:01 crc kubenswrapper[4719]: I1215 12:39:01.885500 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:39:01 crc kubenswrapper[4719]: I1215 12:39:01.898534 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7ff7556d66-bpsdl" event={"ID":"a996feea-1c4e-496f-a91b-2f07b852d22e","Type":"ContainerStarted","Data":"b5ea3d58d5d53c9cfaff05e8ad598abba0547bcfaeaf3bcc617a85a311359104"} Dec 15 12:39:01 crc kubenswrapper[4719]: I1215 12:39:01.898567 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7ff7556d66-bpsdl" event={"ID":"a996feea-1c4e-496f-a91b-2f07b852d22e","Type":"ContainerStarted","Data":"4c1fcd349d122149c8cca4190e24a70adccad8e02cae187240bf9bb38fe3c170"} Dec 15 12:39:01 crc kubenswrapper[4719]: I1215 12:39:01.898581 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:01 crc kubenswrapper[4719]: I1215 12:39:01.898590 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7ff7556d66-bpsdl" event={"ID":"a996feea-1c4e-496f-a91b-2f07b852d22e","Type":"ContainerStarted","Data":"af26bc119af23cb73f61c15829b48382f7d6afe996eb90577a55b2cfbd5a67e5"} Dec 15 12:39:01 crc kubenswrapper[4719]: I1215 12:39:01.899008 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:01 crc kubenswrapper[4719]: I1215 12:39:01.933706 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7ff7556d66-bpsdl" podStartSLOduration=1.933684019 podStartE2EDuration="1.933684019s" podCreationTimestamp="2025-12-15 12:39:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:39:01.931665515 +0000 UTC m=+1302.873958545" watchObservedRunningTime="2025-12-15 12:39:01.933684019 +0000 UTC m=+1302.875977049" Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.352101 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6c86c57db-47pxb" podUID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.395080 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.395178 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.396085 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"a515a01ab9e93d2a61d2349f044b8132b8de12998c6a016d169b104d98e71a08"} pod="openstack/horizon-7c6bcf58b6-vx99l" containerMessage="Container horizon failed startup probe, will be restarted" Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.396136 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" containerID="cri-o://a515a01ab9e93d2a61d2349f044b8132b8de12998c6a016d169b104d98e71a08" gracePeriod=30 Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.581031 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc5fb48-jqf6b" podUID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.581354 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.582075 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"975b276955c100d7fd0352b04978b9b7d4a57b2a3ddba6e0cbb4fd095434fee3"} pod="openstack/horizon-6bc5fb48-jqf6b" containerMessage="Container horizon failed startup probe, will be restarted" Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.582180 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6bc5fb48-jqf6b" podUID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerName="horizon" containerID="cri-o://975b276955c100d7fd0352b04978b9b7d4a57b2a3ddba6e0cbb4fd095434fee3" gracePeriod=30 Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.782953 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 15 12:39:02 crc kubenswrapper[4719]: I1215 12:39:02.866841 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.032011 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.150910 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-52bzg"] Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.151177 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" podUID="46738d80-d21e-43c5-8871-dba878a5986b" containerName="dnsmasq-dns" containerID="cri-o://3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d" gracePeriod=10 Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.714556 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.854076 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-nb\") pod \"46738d80-d21e-43c5-8871-dba878a5986b\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.854214 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-config\") pod \"46738d80-d21e-43c5-8871-dba878a5986b\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.854254 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-sb\") pod \"46738d80-d21e-43c5-8871-dba878a5986b\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.854283 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-svc\") pod \"46738d80-d21e-43c5-8871-dba878a5986b\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.854333 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-swift-storage-0\") pod \"46738d80-d21e-43c5-8871-dba878a5986b\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.854365 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwmwv\" (UniqueName: \"kubernetes.io/projected/46738d80-d21e-43c5-8871-dba878a5986b-kube-api-access-bwmwv\") pod \"46738d80-d21e-43c5-8871-dba878a5986b\" (UID: \"46738d80-d21e-43c5-8871-dba878a5986b\") " Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.905582 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46738d80-d21e-43c5-8871-dba878a5986b-kube-api-access-bwmwv" (OuterVolumeSpecName: "kube-api-access-bwmwv") pod "46738d80-d21e-43c5-8871-dba878a5986b" (UID: "46738d80-d21e-43c5-8871-dba878a5986b"). InnerVolumeSpecName "kube-api-access-bwmwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.958352 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwmwv\" (UniqueName: \"kubernetes.io/projected/46738d80-d21e-43c5-8871-dba878a5986b-kube-api-access-bwmwv\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.989157 4719 generic.go:334] "Generic (PLEG): container finished" podID="46738d80-d21e-43c5-8871-dba878a5986b" containerID="3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d" exitCode=0 Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.989468 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.990367 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" event={"ID":"46738d80-d21e-43c5-8871-dba878a5986b","Type":"ContainerDied","Data":"3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d"} Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.990407 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-52bzg" event={"ID":"46738d80-d21e-43c5-8871-dba878a5986b","Type":"ContainerDied","Data":"3253d1f65a34af251731c5573ca79f11ed3c5d016dee15dc886768ec72c84c78"} Dec 15 12:39:03 crc kubenswrapper[4719]: I1215 12:39:03.990431 4719 scope.go:117] "RemoveContainer" containerID="3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.018550 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "46738d80-d21e-43c5-8871-dba878a5986b" (UID: "46738d80-d21e-43c5-8871-dba878a5986b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.031264 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "46738d80-d21e-43c5-8871-dba878a5986b" (UID: "46738d80-d21e-43c5-8871-dba878a5986b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.039407 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "46738d80-d21e-43c5-8871-dba878a5986b" (UID: "46738d80-d21e-43c5-8871-dba878a5986b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.062986 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-config" (OuterVolumeSpecName: "config") pod "46738d80-d21e-43c5-8871-dba878a5986b" (UID: "46738d80-d21e-43c5-8871-dba878a5986b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.064163 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.064201 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.064215 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.064230 4719 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.070463 4719 scope.go:117] "RemoveContainer" containerID="4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.077110 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "46738d80-d21e-43c5-8871-dba878a5986b" (UID: "46738d80-d21e-43c5-8871-dba878a5986b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.114233 4719 scope.go:117] "RemoveContainer" containerID="3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d" Dec 15 12:39:04 crc kubenswrapper[4719]: E1215 12:39:04.115007 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d\": container with ID starting with 3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d not found: ID does not exist" containerID="3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.115055 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d"} err="failed to get container status \"3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d\": rpc error: code = NotFound desc = could not find container \"3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d\": container with ID starting with 3e9dc1a882fcc0a58dd3f69cfd49607002c78f6a22718a31d7f142995ba3025d not found: ID does not exist" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.115083 4719 scope.go:117] "RemoveContainer" containerID="4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8" Dec 15 12:39:04 crc kubenswrapper[4719]: E1215 12:39:04.115526 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8\": container with ID starting with 4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8 not found: ID does not exist" containerID="4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.115556 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8"} err="failed to get container status \"4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8\": rpc error: code = NotFound desc = could not find container \"4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8\": container with ID starting with 4938d5a4562d83943f34287869ba8215a0c623a5e51ff914da92563f35bbf7d8 not found: ID does not exist" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.167635 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46738d80-d21e-43c5-8871-dba878a5986b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.326787 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-52bzg"] Dec 15 12:39:04 crc kubenswrapper[4719]: I1215 12:39:04.336279 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-52bzg"] Dec 15 12:39:05 crc kubenswrapper[4719]: I1215 12:39:05.488283 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46738d80-d21e-43c5-8871-dba878a5986b" path="/var/lib/kubelet/pods/46738d80-d21e-43c5-8871-dba878a5986b/volumes" Dec 15 12:39:06 crc kubenswrapper[4719]: I1215 12:39:06.512488 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 15 12:39:07 crc kubenswrapper[4719]: I1215 12:39:07.032894 4719 generic.go:334] "Generic (PLEG): container finished" podID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerID="a515a01ab9e93d2a61d2349f044b8132b8de12998c6a016d169b104d98e71a08" exitCode=0 Dec 15 12:39:07 crc kubenswrapper[4719]: I1215 12:39:07.032934 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6bcf58b6-vx99l" event={"ID":"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e","Type":"ContainerDied","Data":"a515a01ab9e93d2a61d2349f044b8132b8de12998c6a016d169b104d98e71a08"} Dec 15 12:39:08 crc kubenswrapper[4719]: I1215 12:39:08.043459 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6bcf58b6-vx99l" event={"ID":"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e","Type":"ContainerStarted","Data":"e526058563f282b4d65e64a01944ac0f829154bf6a70328411350711beca2af4"} Dec 15 12:39:08 crc kubenswrapper[4719]: I1215 12:39:08.628848 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 15 12:39:08 crc kubenswrapper[4719]: I1215 12:39:08.701715 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 15 12:39:09 crc kubenswrapper[4719]: I1215 12:39:09.074948 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:39:09 crc kubenswrapper[4719]: I1215 12:39:09.082311 4719 generic.go:334] "Generic (PLEG): container finished" podID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerID="975b276955c100d7fd0352b04978b9b7d4a57b2a3ddba6e0cbb4fd095434fee3" exitCode=0 Dec 15 12:39:09 crc kubenswrapper[4719]: I1215 12:39:09.083301 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc5fb48-jqf6b" event={"ID":"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926","Type":"ContainerDied","Data":"975b276955c100d7fd0352b04978b9b7d4a57b2a3ddba6e0cbb4fd095434fee3"} Dec 15 12:39:09 crc kubenswrapper[4719]: I1215 12:39:09.083338 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc5fb48-jqf6b" event={"ID":"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926","Type":"ContainerStarted","Data":"b3d3c0d45c63097362f15fdd3a23877d00f86c6a1a144fb9d5b2f5b8da10fbb1"} Dec 15 12:39:09 crc kubenswrapper[4719]: I1215 12:39:09.083830 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5909a62d-a8e1-4a30-bb53-4d68e403f090" containerName="probe" containerID="cri-o://a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142" gracePeriod=30 Dec 15 12:39:09 crc kubenswrapper[4719]: I1215 12:39:09.083791 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5909a62d-a8e1-4a30-bb53-4d68e403f090" containerName="cinder-scheduler" containerID="cri-o://da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b" gracePeriod=30 Dec 15 12:39:09 crc kubenswrapper[4719]: I1215 12:39:09.354058 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-9d9f8bcf4-hj8fd" Dec 15 12:39:09 crc kubenswrapper[4719]: I1215 12:39:09.657531 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5c7789bfff-vvvqf" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.608837 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.723282 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data-custom\") pod \"5909a62d-a8e1-4a30-bb53-4d68e403f090\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.723335 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5909a62d-a8e1-4a30-bb53-4d68e403f090-etc-machine-id\") pod \"5909a62d-a8e1-4a30-bb53-4d68e403f090\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.723364 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-scripts\") pod \"5909a62d-a8e1-4a30-bb53-4d68e403f090\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.723405 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-combined-ca-bundle\") pod \"5909a62d-a8e1-4a30-bb53-4d68e403f090\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.723438 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfmvp\" (UniqueName: \"kubernetes.io/projected/5909a62d-a8e1-4a30-bb53-4d68e403f090-kube-api-access-dfmvp\") pod \"5909a62d-a8e1-4a30-bb53-4d68e403f090\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.723457 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5909a62d-a8e1-4a30-bb53-4d68e403f090-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5909a62d-a8e1-4a30-bb53-4d68e403f090" (UID: "5909a62d-a8e1-4a30-bb53-4d68e403f090"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.723561 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data\") pod \"5909a62d-a8e1-4a30-bb53-4d68e403f090\" (UID: \"5909a62d-a8e1-4a30-bb53-4d68e403f090\") " Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.724753 4719 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5909a62d-a8e1-4a30-bb53-4d68e403f090-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.736132 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5909a62d-a8e1-4a30-bb53-4d68e403f090-kube-api-access-dfmvp" (OuterVolumeSpecName: "kube-api-access-dfmvp") pod "5909a62d-a8e1-4a30-bb53-4d68e403f090" (UID: "5909a62d-a8e1-4a30-bb53-4d68e403f090"). InnerVolumeSpecName "kube-api-access-dfmvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.737017 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5909a62d-a8e1-4a30-bb53-4d68e403f090" (UID: "5909a62d-a8e1-4a30-bb53-4d68e403f090"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.755009 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-scripts" (OuterVolumeSpecName: "scripts") pod "5909a62d-a8e1-4a30-bb53-4d68e403f090" (UID: "5909a62d-a8e1-4a30-bb53-4d68e403f090"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.824832 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data" (OuterVolumeSpecName: "config-data") pod "5909a62d-a8e1-4a30-bb53-4d68e403f090" (UID: "5909a62d-a8e1-4a30-bb53-4d68e403f090"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.826420 4719 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.826451 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.826463 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfmvp\" (UniqueName: \"kubernetes.io/projected/5909a62d-a8e1-4a30-bb53-4d68e403f090-kube-api-access-dfmvp\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.826476 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.853158 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5909a62d-a8e1-4a30-bb53-4d68e403f090" (UID: "5909a62d-a8e1-4a30-bb53-4d68e403f090"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:10 crc kubenswrapper[4719]: I1215 12:39:10.928849 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5909a62d-a8e1-4a30-bb53-4d68e403f090-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.100553 4719 generic.go:334] "Generic (PLEG): container finished" podID="5909a62d-a8e1-4a30-bb53-4d68e403f090" containerID="a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142" exitCode=0 Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.100582 4719 generic.go:334] "Generic (PLEG): container finished" podID="5909a62d-a8e1-4a30-bb53-4d68e403f090" containerID="da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b" exitCode=0 Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.100600 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5909a62d-a8e1-4a30-bb53-4d68e403f090","Type":"ContainerDied","Data":"a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142"} Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.100626 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5909a62d-a8e1-4a30-bb53-4d68e403f090","Type":"ContainerDied","Data":"da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b"} Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.100636 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5909a62d-a8e1-4a30-bb53-4d68e403f090","Type":"ContainerDied","Data":"98df4ad8100837012f11e69a3e57809575174ce33f9c982906943e643d6eb1ec"} Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.100652 4719 scope.go:117] "RemoveContainer" containerID="a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.100761 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.132878 4719 scope.go:117] "RemoveContainer" containerID="da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.143115 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.163636 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.186713 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 15 12:39:11 crc kubenswrapper[4719]: E1215 12:39:11.187108 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46738d80-d21e-43c5-8871-dba878a5986b" containerName="dnsmasq-dns" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.187129 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="46738d80-d21e-43c5-8871-dba878a5986b" containerName="dnsmasq-dns" Dec 15 12:39:11 crc kubenswrapper[4719]: E1215 12:39:11.187167 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46738d80-d21e-43c5-8871-dba878a5986b" containerName="init" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.187175 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="46738d80-d21e-43c5-8871-dba878a5986b" containerName="init" Dec 15 12:39:11 crc kubenswrapper[4719]: E1215 12:39:11.187219 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5909a62d-a8e1-4a30-bb53-4d68e403f090" containerName="probe" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.187226 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5909a62d-a8e1-4a30-bb53-4d68e403f090" containerName="probe" Dec 15 12:39:11 crc kubenswrapper[4719]: E1215 12:39:11.187238 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5909a62d-a8e1-4a30-bb53-4d68e403f090" containerName="cinder-scheduler" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.187244 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5909a62d-a8e1-4a30-bb53-4d68e403f090" containerName="cinder-scheduler" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.190624 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5909a62d-a8e1-4a30-bb53-4d68e403f090" containerName="probe" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.190665 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="46738d80-d21e-43c5-8871-dba878a5986b" containerName="dnsmasq-dns" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.190673 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5909a62d-a8e1-4a30-bb53-4d68e403f090" containerName="cinder-scheduler" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.191767 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.194077 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.207392 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.211106 4719 scope.go:117] "RemoveContainer" containerID="a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142" Dec 15 12:39:11 crc kubenswrapper[4719]: E1215 12:39:11.211633 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142\": container with ID starting with a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142 not found: ID does not exist" containerID="a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.211662 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142"} err="failed to get container status \"a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142\": rpc error: code = NotFound desc = could not find container \"a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142\": container with ID starting with a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142 not found: ID does not exist" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.211683 4719 scope.go:117] "RemoveContainer" containerID="da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b" Dec 15 12:39:11 crc kubenswrapper[4719]: E1215 12:39:11.225269 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b\": container with ID starting with da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b not found: ID does not exist" containerID="da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.225305 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b"} err="failed to get container status \"da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b\": rpc error: code = NotFound desc = could not find container \"da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b\": container with ID starting with da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b not found: ID does not exist" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.225330 4719 scope.go:117] "RemoveContainer" containerID="a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.235125 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142"} err="failed to get container status \"a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142\": rpc error: code = NotFound desc = could not find container \"a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142\": container with ID starting with a6fe913ef86334237c3835adf98e30a9129e84a3da048153267a739db0175142 not found: ID does not exist" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.235188 4719 scope.go:117] "RemoveContainer" containerID="da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.237642 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b"} err="failed to get container status \"da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b\": rpc error: code = NotFound desc = could not find container \"da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b\": container with ID starting with da238cebc94806531c08de44335cbf882aadc4228ec8b36d8746cbcb2d36c32b not found: ID does not exist" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.338700 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.338913 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2p9k\" (UniqueName: \"kubernetes.io/projected/d60e91f3-5bb1-4958-8489-e047f9a31c87-kube-api-access-x2p9k\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.338954 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-scripts\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.339163 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.339213 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d60e91f3-5bb1-4958-8489-e047f9a31c87-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.339288 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-config-data\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.440971 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.441115 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2p9k\" (UniqueName: \"kubernetes.io/projected/d60e91f3-5bb1-4958-8489-e047f9a31c87-kube-api-access-x2p9k\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.441148 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-scripts\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.441283 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.441317 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d60e91f3-5bb1-4958-8489-e047f9a31c87-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.441374 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-config-data\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.441983 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d60e91f3-5bb1-4958-8489-e047f9a31c87-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.448423 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.450518 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-config-data\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.453193 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-scripts\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.469563 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2p9k\" (UniqueName: \"kubernetes.io/projected/d60e91f3-5bb1-4958-8489-e047f9a31c87-kube-api-access-x2p9k\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.478404 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d60e91f3-5bb1-4958-8489-e047f9a31c87-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d60e91f3-5bb1-4958-8489-e047f9a31c87\") " pod="openstack/cinder-scheduler-0" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.488631 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5909a62d-a8e1-4a30-bb53-4d68e403f090" path="/var/lib/kubelet/pods/5909a62d-a8e1-4a30-bb53-4d68e403f090/volumes" Dec 15 12:39:11 crc kubenswrapper[4719]: I1215 12:39:11.536944 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.195948 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.560139 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.791920 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.793641 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.796789 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.803623 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.803798 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-sz982" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.867633 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.878786 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65x4c\" (UniqueName: \"kubernetes.io/projected/0b238728-d6aa-4b27-b48b-beb0cc7ff437-kube-api-access-65x4c\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.879012 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b238728-d6aa-4b27-b48b-beb0cc7ff437-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.879101 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0b238728-d6aa-4b27-b48b-beb0cc7ff437-openstack-config\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.879192 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0b238728-d6aa-4b27-b48b-beb0cc7ff437-openstack-config-secret\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.981070 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65x4c\" (UniqueName: \"kubernetes.io/projected/0b238728-d6aa-4b27-b48b-beb0cc7ff437-kube-api-access-65x4c\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.981533 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b238728-d6aa-4b27-b48b-beb0cc7ff437-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.981638 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0b238728-d6aa-4b27-b48b-beb0cc7ff437-openstack-config\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.981752 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0b238728-d6aa-4b27-b48b-beb0cc7ff437-openstack-config-secret\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.982551 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0b238728-d6aa-4b27-b48b-beb0cc7ff437-openstack-config\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:12 crc kubenswrapper[4719]: I1215 12:39:12.990532 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0b238728-d6aa-4b27-b48b-beb0cc7ff437-openstack-config-secret\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:13 crc kubenswrapper[4719]: I1215 12:39:13.005380 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b238728-d6aa-4b27-b48b-beb0cc7ff437-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:13 crc kubenswrapper[4719]: I1215 12:39:13.008338 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65x4c\" (UniqueName: \"kubernetes.io/projected/0b238728-d6aa-4b27-b48b-beb0cc7ff437-kube-api-access-65x4c\") pod \"openstackclient\" (UID: \"0b238728-d6aa-4b27-b48b-beb0cc7ff437\") " pod="openstack/openstackclient" Dec 15 12:39:13 crc kubenswrapper[4719]: I1215 12:39:13.119685 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d60e91f3-5bb1-4958-8489-e047f9a31c87","Type":"ContainerStarted","Data":"617dce7002f573e60a9b7e298843a0aef12d512da930167c0a00848a649ca4d2"} Dec 15 12:39:13 crc kubenswrapper[4719]: I1215 12:39:13.131396 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 15 12:39:13 crc kubenswrapper[4719]: I1215 12:39:13.700430 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 15 12:39:14 crc kubenswrapper[4719]: I1215 12:39:14.022576 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7ff7556d66-bpsdl" Dec 15 12:39:14 crc kubenswrapper[4719]: I1215 12:39:14.153380 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0b238728-d6aa-4b27-b48b-beb0cc7ff437","Type":"ContainerStarted","Data":"0b8b58e0574258d778a7067ae0fbee17fb7a775a7a1bb9a1d9ddc9d01ac679fd"} Dec 15 12:39:14 crc kubenswrapper[4719]: I1215 12:39:14.159547 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c86c57db-47pxb"] Dec 15 12:39:14 crc kubenswrapper[4719]: I1215 12:39:14.159769 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c86c57db-47pxb" podUID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerName="barbican-api-log" containerID="cri-o://ebff589ae4bfeda370095f36ab298c3c5529a0314fa5beb1b538094a10e05edf" gracePeriod=30 Dec 15 12:39:14 crc kubenswrapper[4719]: I1215 12:39:14.160258 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c86c57db-47pxb" podUID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerName="barbican-api" containerID="cri-o://28782885e6f9c77a44600262069f881d5a034ddbee91fa23088db779f8574dfe" gracePeriod=30 Dec 15 12:39:14 crc kubenswrapper[4719]: I1215 12:39:14.179101 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d60e91f3-5bb1-4958-8489-e047f9a31c87","Type":"ContainerStarted","Data":"d2a72584c15a752292ab9038083a84216bdd061be01aa3e210e0d77b9e56bea0"} Dec 15 12:39:15 crc kubenswrapper[4719]: I1215 12:39:15.188584 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d60e91f3-5bb1-4958-8489-e047f9a31c87","Type":"ContainerStarted","Data":"286fa5d5f0a8b94c77000d8bb2c4e3f8b07b3e26a7a7fd9863b1efed6209eed5"} Dec 15 12:39:15 crc kubenswrapper[4719]: I1215 12:39:15.191198 4719 generic.go:334] "Generic (PLEG): container finished" podID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerID="ebff589ae4bfeda370095f36ab298c3c5529a0314fa5beb1b538094a10e05edf" exitCode=143 Dec 15 12:39:15 crc kubenswrapper[4719]: I1215 12:39:15.191280 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86c57db-47pxb" event={"ID":"7818f1dd-1db1-4af7-b833-e4bbc91e9a09","Type":"ContainerDied","Data":"ebff589ae4bfeda370095f36ab298c3c5529a0314fa5beb1b538094a10e05edf"} Dec 15 12:39:15 crc kubenswrapper[4719]: I1215 12:39:15.212211 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.212188155 podStartE2EDuration="4.212188155s" podCreationTimestamp="2025-12-15 12:39:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:39:15.204584046 +0000 UTC m=+1316.146877076" watchObservedRunningTime="2025-12-15 12:39:15.212188155 +0000 UTC m=+1316.154481195" Dec 15 12:39:16 crc kubenswrapper[4719]: I1215 12:39:16.537461 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 15 12:39:17 crc kubenswrapper[4719]: I1215 12:39:17.388889 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:39:17 crc kubenswrapper[4719]: I1215 12:39:17.389217 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:39:17 crc kubenswrapper[4719]: I1215 12:39:17.390266 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 15 12:39:17 crc kubenswrapper[4719]: I1215 12:39:17.496848 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 15 12:39:17 crc kubenswrapper[4719]: I1215 12:39:17.575634 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:39:17 crc kubenswrapper[4719]: I1215 12:39:17.575671 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.182013 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.163:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.220815 4719 generic.go:334] "Generic (PLEG): container finished" podID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerID="28782885e6f9c77a44600262069f881d5a034ddbee91fa23088db779f8574dfe" exitCode=0 Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.220856 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86c57db-47pxb" event={"ID":"7818f1dd-1db1-4af7-b833-e4bbc91e9a09","Type":"ContainerDied","Data":"28782885e6f9c77a44600262069f881d5a034ddbee91fa23088db779f8574dfe"} Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.303029 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.388797 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjtcn\" (UniqueName: \"kubernetes.io/projected/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-kube-api-access-rjtcn\") pod \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.388982 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-logs\") pod \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.389018 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data-custom\") pod \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.389097 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-combined-ca-bundle\") pod \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.389240 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data\") pod \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\" (UID: \"7818f1dd-1db1-4af7-b833-e4bbc91e9a09\") " Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.391057 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-logs" (OuterVolumeSpecName: "logs") pod "7818f1dd-1db1-4af7-b833-e4bbc91e9a09" (UID: "7818f1dd-1db1-4af7-b833-e4bbc91e9a09"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.415156 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-kube-api-access-rjtcn" (OuterVolumeSpecName: "kube-api-access-rjtcn") pod "7818f1dd-1db1-4af7-b833-e4bbc91e9a09" (UID: "7818f1dd-1db1-4af7-b833-e4bbc91e9a09"). InnerVolumeSpecName "kube-api-access-rjtcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.417018 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7818f1dd-1db1-4af7-b833-e4bbc91e9a09" (UID: "7818f1dd-1db1-4af7-b833-e4bbc91e9a09"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.465474 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7818f1dd-1db1-4af7-b833-e4bbc91e9a09" (UID: "7818f1dd-1db1-4af7-b833-e4bbc91e9a09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.478500 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data" (OuterVolumeSpecName: "config-data") pod "7818f1dd-1db1-4af7-b833-e4bbc91e9a09" (UID: "7818f1dd-1db1-4af7-b833-e4bbc91e9a09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.493186 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjtcn\" (UniqueName: \"kubernetes.io/projected/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-kube-api-access-rjtcn\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.493460 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.493553 4719 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.493629 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:18 crc kubenswrapper[4719]: I1215 12:39:18.493700 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7818f1dd-1db1-4af7-b833-e4bbc91e9a09-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:19 crc kubenswrapper[4719]: I1215 12:39:19.236999 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c86c57db-47pxb" event={"ID":"7818f1dd-1db1-4af7-b833-e4bbc91e9a09","Type":"ContainerDied","Data":"ff02c273660cef618806136d485846099b778e88c40235e63f0a6b6cc291b13d"} Dec 15 12:39:19 crc kubenswrapper[4719]: I1215 12:39:19.237263 4719 scope.go:117] "RemoveContainer" containerID="28782885e6f9c77a44600262069f881d5a034ddbee91fa23088db779f8574dfe" Dec 15 12:39:19 crc kubenswrapper[4719]: I1215 12:39:19.237386 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c86c57db-47pxb" Dec 15 12:39:19 crc kubenswrapper[4719]: I1215 12:39:19.263311 4719 scope.go:117] "RemoveContainer" containerID="ebff589ae4bfeda370095f36ab298c3c5529a0314fa5beb1b538094a10e05edf" Dec 15 12:39:19 crc kubenswrapper[4719]: I1215 12:39:19.304947 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c86c57db-47pxb"] Dec 15 12:39:19 crc kubenswrapper[4719]: I1215 12:39:19.346176 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6c86c57db-47pxb"] Dec 15 12:39:19 crc kubenswrapper[4719]: E1215 12:39:19.348709 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7818f1dd_1db1_4af7_b833_e4bbc91e9a09.slice/crio-ff02c273660cef618806136d485846099b778e88c40235e63f0a6b6cc291b13d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7818f1dd_1db1_4af7_b833_e4bbc91e9a09.slice\": RecentStats: unable to find data in memory cache]" Dec 15 12:39:19 crc kubenswrapper[4719]: I1215 12:39:19.496447 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" path="/var/lib/kubelet/pods/7818f1dd-1db1-4af7-b833-e4bbc91e9a09/volumes" Dec 15 12:39:19 crc kubenswrapper[4719]: I1215 12:39:19.936625 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5649967964-ctb22" Dec 15 12:39:20 crc kubenswrapper[4719]: I1215 12:39:20.774544 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:20 crc kubenswrapper[4719]: I1215 12:39:20.774886 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="sg-core" containerID="cri-o://87aabd4d0941940cd26018e043a9abce5ca108f30d47e257c61cb6570c697f40" gracePeriod=30 Dec 15 12:39:20 crc kubenswrapper[4719]: I1215 12:39:20.775010 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="proxy-httpd" containerID="cri-o://a3a301cb98da46f13ca27360ca1c76299fafc4f89aa2582e83f52a2477149de9" gracePeriod=30 Dec 15 12:39:20 crc kubenswrapper[4719]: I1215 12:39:20.775046 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="ceilometer-notification-agent" containerID="cri-o://3cb74a676d33c8c35242627d7a2d15a9dab16f3673e70b096107042b68e7395c" gracePeriod=30 Dec 15 12:39:20 crc kubenswrapper[4719]: I1215 12:39:20.774842 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="ceilometer-central-agent" containerID="cri-o://d67ccc630a371b702f7cba088063782b8442732e4ad34971ee7482ddb314d390" gracePeriod=30 Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.214403 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7f6dc9444c-2jcmv"] Dec 15 12:39:21 crc kubenswrapper[4719]: E1215 12:39:21.214879 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerName="barbican-api" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.214901 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerName="barbican-api" Dec 15 12:39:21 crc kubenswrapper[4719]: E1215 12:39:21.214925 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerName="barbican-api-log" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.214933 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerName="barbican-api-log" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.215176 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerName="barbican-api-log" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.215207 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="7818f1dd-1db1-4af7-b833-e4bbc91e9a09" containerName="barbican-api" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.231160 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.234103 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.251195 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.251291 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/60352f3a-8ee6-498d-9ee3-152cc4ab562f-etc-swift\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.251332 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z589h\" (UniqueName: \"kubernetes.io/projected/60352f3a-8ee6-498d-9ee3-152cc4ab562f-kube-api-access-z589h\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.251358 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-combined-ca-bundle\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.251433 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60352f3a-8ee6-498d-9ee3-152cc4ab562f-run-httpd\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.251453 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.251454 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-internal-tls-certs\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.259978 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-config-data\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.260030 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60352f3a-8ee6-498d-9ee3-152cc4ab562f-log-httpd\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.260137 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-public-tls-certs\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.299640 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7f6dc9444c-2jcmv"] Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.308036 4719 generic.go:334] "Generic (PLEG): container finished" podID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerID="a3a301cb98da46f13ca27360ca1c76299fafc4f89aa2582e83f52a2477149de9" exitCode=0 Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.308068 4719 generic.go:334] "Generic (PLEG): container finished" podID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerID="87aabd4d0941940cd26018e043a9abce5ca108f30d47e257c61cb6570c697f40" exitCode=2 Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.308087 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf7e6ad7-6983-4455-af5d-5f7dda131dd2","Type":"ContainerDied","Data":"a3a301cb98da46f13ca27360ca1c76299fafc4f89aa2582e83f52a2477149de9"} Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.308109 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf7e6ad7-6983-4455-af5d-5f7dda131dd2","Type":"ContainerDied","Data":"87aabd4d0941940cd26018e043a9abce5ca108f30d47e257c61cb6570c697f40"} Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.356434 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.356483 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.367268 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60352f3a-8ee6-498d-9ee3-152cc4ab562f-run-httpd\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.367309 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-internal-tls-certs\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.367325 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-config-data\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.367341 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60352f3a-8ee6-498d-9ee3-152cc4ab562f-log-httpd\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.367381 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-public-tls-certs\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.367441 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/60352f3a-8ee6-498d-9ee3-152cc4ab562f-etc-swift\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.367462 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z589h\" (UniqueName: \"kubernetes.io/projected/60352f3a-8ee6-498d-9ee3-152cc4ab562f-kube-api-access-z589h\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.367478 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-combined-ca-bundle\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.368041 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60352f3a-8ee6-498d-9ee3-152cc4ab562f-log-httpd\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.368266 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60352f3a-8ee6-498d-9ee3-152cc4ab562f-run-httpd\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.375650 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-combined-ca-bundle\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.376840 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-public-tls-certs\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.377131 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/60352f3a-8ee6-498d-9ee3-152cc4ab562f-etc-swift\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.377639 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-internal-tls-certs\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.381871 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60352f3a-8ee6-498d-9ee3-152cc4ab562f-config-data\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.396458 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z589h\" (UniqueName: \"kubernetes.io/projected/60352f3a-8ee6-498d-9ee3-152cc4ab562f-kube-api-access-z589h\") pod \"swift-proxy-7f6dc9444c-2jcmv\" (UID: \"60352f3a-8ee6-498d-9ee3-152cc4ab562f\") " pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.586927 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:21 crc kubenswrapper[4719]: I1215 12:39:21.840295 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 15 12:39:22 crc kubenswrapper[4719]: I1215 12:39:22.324204 4719 generic.go:334] "Generic (PLEG): container finished" podID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerID="d67ccc630a371b702f7cba088063782b8442732e4ad34971ee7482ddb314d390" exitCode=0 Dec 15 12:39:22 crc kubenswrapper[4719]: I1215 12:39:22.324252 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf7e6ad7-6983-4455-af5d-5f7dda131dd2","Type":"ContainerDied","Data":"d67ccc630a371b702f7cba088063782b8442732e4ad34971ee7482ddb314d390"} Dec 15 12:39:23 crc kubenswrapper[4719]: I1215 12:39:23.334278 4719 generic.go:334] "Generic (PLEG): container finished" podID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerID="3cb74a676d33c8c35242627d7a2d15a9dab16f3673e70b096107042b68e7395c" exitCode=0 Dec 15 12:39:23 crc kubenswrapper[4719]: I1215 12:39:23.334328 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf7e6ad7-6983-4455-af5d-5f7dda131dd2","Type":"ContainerDied","Data":"3cb74a676d33c8c35242627d7a2d15a9dab16f3673e70b096107042b68e7395c"} Dec 15 12:39:24 crc kubenswrapper[4719]: I1215 12:39:24.929541 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:39:24 crc kubenswrapper[4719]: I1215 12:39:24.930162 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="51211ffc-0363-4e48-9972-15b8e997d4de" containerName="glance-log" containerID="cri-o://2f5fc5d9ab5592e8a600bbf12061874b624c5e60975cf97b442cd6354eacc609" gracePeriod=30 Dec 15 12:39:24 crc kubenswrapper[4719]: I1215 12:39:24.930331 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="51211ffc-0363-4e48-9972-15b8e997d4de" containerName="glance-httpd" containerID="cri-o://04f9a53ed870a660d04118a7bc49f93fd4196e4150a668f6be6bcc5b0b375c32" gracePeriod=30 Dec 15 12:39:25 crc kubenswrapper[4719]: I1215 12:39:25.355250 4719 generic.go:334] "Generic (PLEG): container finished" podID="51211ffc-0363-4e48-9972-15b8e997d4de" containerID="2f5fc5d9ab5592e8a600bbf12061874b624c5e60975cf97b442cd6354eacc609" exitCode=143 Dec 15 12:39:25 crc kubenswrapper[4719]: I1215 12:39:25.355355 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51211ffc-0363-4e48-9972-15b8e997d4de","Type":"ContainerDied","Data":"2f5fc5d9ab5592e8a600bbf12061874b624c5e60975cf97b442cd6354eacc609"} Dec 15 12:39:27 crc kubenswrapper[4719]: I1215 12:39:27.389419 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 15 12:39:27 crc kubenswrapper[4719]: I1215 12:39:27.577588 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc5fb48-jqf6b" podUID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 15 12:39:27 crc kubenswrapper[4719]: I1215 12:39:27.982085 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:39:27 crc kubenswrapper[4719]: I1215 12:39:27.982727 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="63366e84-551e-4eb0-a007-7156b6ab1d6b" containerName="glance-httpd" containerID="cri-o://be535f9264accf5731790a1bfd0b7a52d63b8830d8427c919804c75fedbd7290" gracePeriod=30 Dec 15 12:39:27 crc kubenswrapper[4719]: I1215 12:39:27.982363 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="63366e84-551e-4eb0-a007-7156b6ab1d6b" containerName="glance-log" containerID="cri-o://ac4e16b1d34616e8a23c7400e0474377f48d038a2b38a39b0094520b9e4d1a07" gracePeriod=30 Dec 15 12:39:28 crc kubenswrapper[4719]: I1215 12:39:28.385298 4719 generic.go:334] "Generic (PLEG): container finished" podID="51211ffc-0363-4e48-9972-15b8e997d4de" containerID="04f9a53ed870a660d04118a7bc49f93fd4196e4150a668f6be6bcc5b0b375c32" exitCode=0 Dec 15 12:39:28 crc kubenswrapper[4719]: I1215 12:39:28.385382 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51211ffc-0363-4e48-9972-15b8e997d4de","Type":"ContainerDied","Data":"04f9a53ed870a660d04118a7bc49f93fd4196e4150a668f6be6bcc5b0b375c32"} Dec 15 12:39:28 crc kubenswrapper[4719]: I1215 12:39:28.388459 4719 generic.go:334] "Generic (PLEG): container finished" podID="63366e84-551e-4eb0-a007-7156b6ab1d6b" containerID="ac4e16b1d34616e8a23c7400e0474377f48d038a2b38a39b0094520b9e4d1a07" exitCode=143 Dec 15 12:39:28 crc kubenswrapper[4719]: I1215 12:39:28.388494 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63366e84-551e-4eb0-a007-7156b6ab1d6b","Type":"ContainerDied","Data":"ac4e16b1d34616e8a23c7400e0474377f48d038a2b38a39b0094520b9e4d1a07"} Dec 15 12:39:28 crc kubenswrapper[4719]: I1215 12:39:28.601350 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5668b96c97-48pkv" Dec 15 12:39:28 crc kubenswrapper[4719]: I1215 12:39:28.686745 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5649967964-ctb22"] Dec 15 12:39:28 crc kubenswrapper[4719]: I1215 12:39:28.687282 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5649967964-ctb22" podUID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" containerName="neutron-api" containerID="cri-o://6c8f61185a298bc47415e2490728b2af74efe149d868a8e1c30bbe335d4b351b" gracePeriod=30 Dec 15 12:39:28 crc kubenswrapper[4719]: I1215 12:39:28.687678 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5649967964-ctb22" podUID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" containerName="neutron-httpd" containerID="cri-o://594ed6de5e5587a3240d090ec7dcad0447ce09468e5b92aef60bbc735df86854" gracePeriod=30 Dec 15 12:39:29 crc kubenswrapper[4719]: I1215 12:39:29.419268 4719 generic.go:334] "Generic (PLEG): container finished" podID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" containerID="594ed6de5e5587a3240d090ec7dcad0447ce09468e5b92aef60bbc735df86854" exitCode=0 Dec 15 12:39:29 crc kubenswrapper[4719]: I1215 12:39:29.419354 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5649967964-ctb22" event={"ID":"2a1fc07a-9062-4ed7-8610-065fa84f1b2b","Type":"ContainerDied","Data":"594ed6de5e5587a3240d090ec7dcad0447ce09468e5b92aef60bbc735df86854"} Dec 15 12:39:29 crc kubenswrapper[4719]: I1215 12:39:29.422147 4719 generic.go:334] "Generic (PLEG): container finished" podID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerID="03a72fcca616f5e0c4506044788aeee363458eaccd166093db389ab297da0d56" exitCode=137 Dec 15 12:39:29 crc kubenswrapper[4719]: I1215 12:39:29.422181 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adf32d03-4559-4d49-8577-c0dac00c1fec","Type":"ContainerDied","Data":"03a72fcca616f5e0c4506044788aeee363458eaccd166093db389ab297da0d56"} Dec 15 12:39:31 crc kubenswrapper[4719]: E1215 12:39:31.587127 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Dec 15 12:39:31 crc kubenswrapper[4719]: E1215 12:39:31.589396 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n98h5f6h67fhfch658h684h65fh7fh656h696h98hb9h579hdch684hf7h5b9h58bh684h75h694h5bh5fh56dhb8h647h65dhcbh59ch5dh5bfh597q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-65x4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(0b238728-d6aa-4b27-b48b-beb0cc7ff437): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 12:39:31 crc kubenswrapper[4719]: E1215 12:39:31.591232 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="0b238728-d6aa-4b27-b48b-beb0cc7ff437" Dec 15 12:39:31 crc kubenswrapper[4719]: I1215 12:39:31.900458 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7f6dc9444c-2jcmv"] Dec 15 12:39:31 crc kubenswrapper[4719]: I1215 12:39:31.997441 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.027404 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099336 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-scripts\") pod \"51211ffc-0363-4e48-9972-15b8e997d4de\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099373 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-logs\") pod \"51211ffc-0363-4e48-9972-15b8e997d4de\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099391 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-internal-tls-certs\") pod \"51211ffc-0363-4e48-9972-15b8e997d4de\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099416 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dghkz\" (UniqueName: \"kubernetes.io/projected/51211ffc-0363-4e48-9972-15b8e997d4de-kube-api-access-dghkz\") pod \"51211ffc-0363-4e48-9972-15b8e997d4de\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099437 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-config-data\") pod \"51211ffc-0363-4e48-9972-15b8e997d4de\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099465 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-scripts\") pod \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099482 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-config-data\") pod \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099503 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-combined-ca-bundle\") pod \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099566 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnwhp\" (UniqueName: \"kubernetes.io/projected/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-kube-api-access-dnwhp\") pod \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099597 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-combined-ca-bundle\") pod \"51211ffc-0363-4e48-9972-15b8e997d4de\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099615 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"51211ffc-0363-4e48-9972-15b8e997d4de\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099633 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-log-httpd\") pod \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099651 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-httpd-run\") pod \"51211ffc-0363-4e48-9972-15b8e997d4de\" (UID: \"51211ffc-0363-4e48-9972-15b8e997d4de\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099671 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-sg-core-conf-yaml\") pod \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.099688 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-run-httpd\") pod \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\" (UID: \"cf7e6ad7-6983-4455-af5d-5f7dda131dd2\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.114196 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-logs" (OuterVolumeSpecName: "logs") pod "51211ffc-0363-4e48-9972-15b8e997d4de" (UID: "51211ffc-0363-4e48-9972-15b8e997d4de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.127104 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cf7e6ad7-6983-4455-af5d-5f7dda131dd2" (UID: "cf7e6ad7-6983-4455-af5d-5f7dda131dd2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.127229 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-scripts" (OuterVolumeSpecName: "scripts") pod "51211ffc-0363-4e48-9972-15b8e997d4de" (UID: "51211ffc-0363-4e48-9972-15b8e997d4de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.131351 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "51211ffc-0363-4e48-9972-15b8e997d4de" (UID: "51211ffc-0363-4e48-9972-15b8e997d4de"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.131755 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cf7e6ad7-6983-4455-af5d-5f7dda131dd2" (UID: "cf7e6ad7-6983-4455-af5d-5f7dda131dd2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.134827 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-kube-api-access-dnwhp" (OuterVolumeSpecName: "kube-api-access-dnwhp") pod "cf7e6ad7-6983-4455-af5d-5f7dda131dd2" (UID: "cf7e6ad7-6983-4455-af5d-5f7dda131dd2"). InnerVolumeSpecName "kube-api-access-dnwhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.153710 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-scripts" (OuterVolumeSpecName: "scripts") pod "cf7e6ad7-6983-4455-af5d-5f7dda131dd2" (UID: "cf7e6ad7-6983-4455-af5d-5f7dda131dd2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.156035 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "51211ffc-0363-4e48-9972-15b8e997d4de" (UID: "51211ffc-0363-4e48-9972-15b8e997d4de"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.156179 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51211ffc-0363-4e48-9972-15b8e997d4de-kube-api-access-dghkz" (OuterVolumeSpecName: "kube-api-access-dghkz") pod "51211ffc-0363-4e48-9972-15b8e997d4de" (UID: "51211ffc-0363-4e48-9972-15b8e997d4de"). InnerVolumeSpecName "kube-api-access-dghkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.179786 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"51211ffc-0363-4e48-9972-15b8e997d4de","Type":"ContainerDied","Data":"b89c621b3f445671a6352d562d6278fc19f7474fd4d123453295abaaf25e9f8f"} Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.181968 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.179847 4719 scope.go:117] "RemoveContainer" containerID="04f9a53ed870a660d04118a7bc49f93fd4196e4150a668f6be6bcc5b0b375c32" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.202386 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.202424 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnwhp\" (UniqueName: \"kubernetes.io/projected/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-kube-api-access-dnwhp\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.202468 4719 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.202479 4719 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.202490 4719 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.202499 4719 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.202528 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.202536 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51211ffc-0363-4e48-9972-15b8e997d4de-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.202544 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dghkz\" (UniqueName: \"kubernetes.io/projected/51211ffc-0363-4e48-9972-15b8e997d4de-kube-api-access-dghkz\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.210079 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7f6dc9444c-2jcmv" event={"ID":"60352f3a-8ee6-498d-9ee3-152cc4ab562f","Type":"ContainerStarted","Data":"80de25da21df75d83c669b28ad84843ce70edbb792cf4d82671fc61ccb4aafa5"} Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.227162 4719 generic.go:334] "Generic (PLEG): container finished" podID="63366e84-551e-4eb0-a007-7156b6ab1d6b" containerID="be535f9264accf5731790a1bfd0b7a52d63b8830d8427c919804c75fedbd7290" exitCode=0 Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.227245 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63366e84-551e-4eb0-a007-7156b6ab1d6b","Type":"ContainerDied","Data":"be535f9264accf5731790a1bfd0b7a52d63b8830d8427c919804c75fedbd7290"} Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.247019 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.247320 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cf7e6ad7-6983-4455-af5d-5f7dda131dd2","Type":"ContainerDied","Data":"e2c91729a53eae89759a916dd91f234e3df5233e9eb0be468093f37b8fc66fb0"} Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.264006 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="0b238728-d6aa-4b27-b48b-beb0cc7ff437" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.332740 4719 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.362734 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cf7e6ad7-6983-4455-af5d-5f7dda131dd2" (UID: "cf7e6ad7-6983-4455-af5d-5f7dda131dd2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.399656 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51211ffc-0363-4e48-9972-15b8e997d4de" (UID: "51211ffc-0363-4e48-9972-15b8e997d4de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.414217 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.414252 4719 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.414264 4719 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.438155 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "51211ffc-0363-4e48-9972-15b8e997d4de" (UID: "51211ffc-0363-4e48-9972-15b8e997d4de"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.469631 4719 scope.go:117] "RemoveContainer" containerID="2f5fc5d9ab5592e8a600bbf12061874b624c5e60975cf97b442cd6354eacc609" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.473360 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.475036 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-config-data" (OuterVolumeSpecName: "config-data") pod "51211ffc-0363-4e48-9972-15b8e997d4de" (UID: "51211ffc-0363-4e48-9972-15b8e997d4de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.514727 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf7e6ad7-6983-4455-af5d-5f7dda131dd2" (UID: "cf7e6ad7-6983-4455-af5d-5f7dda131dd2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.515827 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.516667 4719 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.516686 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51211ffc-0363-4e48-9972-15b8e997d4de-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.527728 4719 scope.go:117] "RemoveContainer" containerID="a3a301cb98da46f13ca27360ca1c76299fafc4f89aa2582e83f52a2477149de9" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.552090 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.570618 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.584177 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-config-data" (OuterVolumeSpecName: "config-data") pod "cf7e6ad7-6983-4455-af5d-5f7dda131dd2" (UID: "cf7e6ad7-6983-4455-af5d-5f7dda131dd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.617478 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftjtb\" (UniqueName: \"kubernetes.io/projected/adf32d03-4559-4d49-8577-c0dac00c1fec-kube-api-access-ftjtb\") pod \"adf32d03-4559-4d49-8577-c0dac00c1fec\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.617526 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data\") pod \"adf32d03-4559-4d49-8577-c0dac00c1fec\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.617577 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-scripts\") pod \"adf32d03-4559-4d49-8577-c0dac00c1fec\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.617666 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data-custom\") pod \"adf32d03-4559-4d49-8577-c0dac00c1fec\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.617715 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-combined-ca-bundle\") pod \"adf32d03-4559-4d49-8577-c0dac00c1fec\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.617791 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adf32d03-4559-4d49-8577-c0dac00c1fec-etc-machine-id\") pod \"adf32d03-4559-4d49-8577-c0dac00c1fec\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.617842 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adf32d03-4559-4d49-8577-c0dac00c1fec-logs\") pod \"adf32d03-4559-4d49-8577-c0dac00c1fec\" (UID: \"adf32d03-4559-4d49-8577-c0dac00c1fec\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.618783 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/adf32d03-4559-4d49-8577-c0dac00c1fec-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "adf32d03-4559-4d49-8577-c0dac00c1fec" (UID: "adf32d03-4559-4d49-8577-c0dac00c1fec"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.618800 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adf32d03-4559-4d49-8577-c0dac00c1fec-logs" (OuterVolumeSpecName: "logs") pod "adf32d03-4559-4d49-8577-c0dac00c1fec" (UID: "adf32d03-4559-4d49-8577-c0dac00c1fec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.619059 4719 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/adf32d03-4559-4d49-8577-c0dac00c1fec-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.619074 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf7e6ad7-6983-4455-af5d-5f7dda131dd2-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.619082 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adf32d03-4559-4d49-8577-c0dac00c1fec-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.622036 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-scripts" (OuterVolumeSpecName: "scripts") pod "adf32d03-4559-4d49-8577-c0dac00c1fec" (UID: "adf32d03-4559-4d49-8577-c0dac00c1fec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.622186 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "adf32d03-4559-4d49-8577-c0dac00c1fec" (UID: "adf32d03-4559-4d49-8577-c0dac00c1fec"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.622333 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adf32d03-4559-4d49-8577-c0dac00c1fec-kube-api-access-ftjtb" (OuterVolumeSpecName: "kube-api-access-ftjtb") pod "adf32d03-4559-4d49-8577-c0dac00c1fec" (UID: "adf32d03-4559-4d49-8577-c0dac00c1fec"). InnerVolumeSpecName "kube-api-access-ftjtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.649582 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adf32d03-4559-4d49-8577-c0dac00c1fec" (UID: "adf32d03-4559-4d49-8577-c0dac00c1fec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.692056 4719 scope.go:117] "RemoveContainer" containerID="87aabd4d0941940cd26018e043a9abce5ca108f30d47e257c61cb6570c697f40" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.730930 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.731316 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51211ffc-0363-4e48-9972-15b8e997d4de" containerName="glance-log" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731352 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="51211ffc-0363-4e48-9972-15b8e997d4de" containerName="glance-log" Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.731361 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="ceilometer-central-agent" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731370 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="ceilometer-central-agent" Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.731382 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="ceilometer-notification-agent" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731388 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="ceilometer-notification-agent" Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.731412 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerName="cinder-api-log" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731418 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerName="cinder-api-log" Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.731425 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51211ffc-0363-4e48-9972-15b8e997d4de" containerName="glance-httpd" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731431 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="51211ffc-0363-4e48-9972-15b8e997d4de" containerName="glance-httpd" Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.731446 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerName="cinder-api" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731452 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerName="cinder-api" Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.731461 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="sg-core" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731468 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="sg-core" Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.731481 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="proxy-httpd" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731489 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="proxy-httpd" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731658 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="ceilometer-central-agent" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731670 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerName="cinder-api" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731682 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="51211ffc-0363-4e48-9972-15b8e997d4de" containerName="glance-log" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731695 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="sg-core" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731706 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="adf32d03-4559-4d49-8577-c0dac00c1fec" containerName="cinder-api-log" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731717 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="proxy-httpd" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731724 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" containerName="ceilometer-notification-agent" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.731737 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="51211ffc-0363-4e48-9972-15b8e997d4de" containerName="glance-httpd" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.732701 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.740188 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.741412 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftjtb\" (UniqueName: \"kubernetes.io/projected/adf32d03-4559-4d49-8577-c0dac00c1fec-kube-api-access-ftjtb\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.741460 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.741470 4719 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.741480 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.742842 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.773686 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.775329 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data" (OuterVolumeSpecName: "config-data") pod "adf32d03-4559-4d49-8577-c0dac00c1fec" (UID: "adf32d03-4559-4d49-8577-c0dac00c1fec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.814375 4719 scope.go:117] "RemoveContainer" containerID="3cb74a676d33c8c35242627d7a2d15a9dab16f3673e70b096107042b68e7395c" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.845012 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.845080 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.845158 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.845218 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f26a2026-ce1a-4411-8e19-e1981be6f927-logs\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.845247 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfqdx\" (UniqueName: \"kubernetes.io/projected/f26a2026-ce1a-4411-8e19-e1981be6f927-kube-api-access-zfqdx\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.845293 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.845316 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.845504 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f26a2026-ce1a-4411-8e19-e1981be6f927-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.845626 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adf32d03-4559-4d49-8577-c0dac00c1fec-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.846387 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.881614 4719 scope.go:117] "RemoveContainer" containerID="d67ccc630a371b702f7cba088063782b8442732e4ad34971ee7482ddb314d390" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.920156 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.922840 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.947244 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbj2l\" (UniqueName: \"kubernetes.io/projected/63366e84-551e-4eb0-a007-7156b6ab1d6b-kube-api-access-tbj2l\") pod \"63366e84-551e-4eb0-a007-7156b6ab1d6b\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.947341 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-public-tls-certs\") pod \"63366e84-551e-4eb0-a007-7156b6ab1d6b\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.947421 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-httpd-run\") pod \"63366e84-551e-4eb0-a007-7156b6ab1d6b\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.947477 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-scripts\") pod \"63366e84-551e-4eb0-a007-7156b6ab1d6b\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.947525 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"63366e84-551e-4eb0-a007-7156b6ab1d6b\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.947604 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-config-data\") pod \"63366e84-551e-4eb0-a007-7156b6ab1d6b\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.947638 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-combined-ca-bundle\") pod \"63366e84-551e-4eb0-a007-7156b6ab1d6b\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.947685 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-logs\") pod \"63366e84-551e-4eb0-a007-7156b6ab1d6b\" (UID: \"63366e84-551e-4eb0-a007-7156b6ab1d6b\") " Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.947921 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f26a2026-ce1a-4411-8e19-e1981be6f927-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.948035 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.948079 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.948124 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.948154 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f26a2026-ce1a-4411-8e19-e1981be6f927-logs\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.948179 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfqdx\" (UniqueName: \"kubernetes.io/projected/f26a2026-ce1a-4411-8e19-e1981be6f927-kube-api-access-zfqdx\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.948203 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.948225 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.951468 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "63366e84-551e-4eb0-a007-7156b6ab1d6b" (UID: "63366e84-551e-4eb0-a007-7156b6ab1d6b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.952079 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-logs" (OuterVolumeSpecName: "logs") pod "63366e84-551e-4eb0-a007-7156b6ab1d6b" (UID: "63366e84-551e-4eb0-a007-7156b6ab1d6b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.958414 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f26a2026-ce1a-4411-8e19-e1981be6f927-logs\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.967260 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.968879 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63366e84-551e-4eb0-a007-7156b6ab1d6b" containerName="glance-httpd" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.968985 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="63366e84-551e-4eb0-a007-7156b6ab1d6b" containerName="glance-httpd" Dec 15 12:39:32 crc kubenswrapper[4719]: E1215 12:39:32.969076 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63366e84-551e-4eb0-a007-7156b6ab1d6b" containerName="glance-log" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.969130 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="63366e84-551e-4eb0-a007-7156b6ab1d6b" containerName="glance-log" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.969357 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="63366e84-551e-4eb0-a007-7156b6ab1d6b" containerName="glance-log" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.969430 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="63366e84-551e-4eb0-a007-7156b6ab1d6b" containerName="glance-httpd" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.971727 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.972961 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f26a2026-ce1a-4411-8e19-e1981be6f927-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.973513 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.995499 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 15 12:39:32 crc kubenswrapper[4719]: I1215 12:39:32.995781 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.017838 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.042691 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfqdx\" (UniqueName: \"kubernetes.io/projected/f26a2026-ce1a-4411-8e19-e1981be6f927-kube-api-access-zfqdx\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.053277 4719 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.053321 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63366e84-551e-4eb0-a007-7156b6ab1d6b-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.058837 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-scripts" (OuterVolumeSpecName: "scripts") pod "63366e84-551e-4eb0-a007-7156b6ab1d6b" (UID: "63366e84-551e-4eb0-a007-7156b6ab1d6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.068117 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "63366e84-551e-4eb0-a007-7156b6ab1d6b" (UID: "63366e84-551e-4eb0-a007-7156b6ab1d6b"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.068988 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.075246 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63366e84-551e-4eb0-a007-7156b6ab1d6b" (UID: "63366e84-551e-4eb0-a007-7156b6ab1d6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.075361 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63366e84-551e-4eb0-a007-7156b6ab1d6b-kube-api-access-tbj2l" (OuterVolumeSpecName: "kube-api-access-tbj2l") pod "63366e84-551e-4eb0-a007-7156b6ab1d6b" (UID: "63366e84-551e-4eb0-a007-7156b6ab1d6b"). InnerVolumeSpecName "kube-api-access-tbj2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.076617 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.105763 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.122971 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f26a2026-ce1a-4411-8e19-e1981be6f927-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.123656 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f26a2026-ce1a-4411-8e19-e1981be6f927\") " pod="openstack/glance-default-internal-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.146144 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-config-data" (OuterVolumeSpecName: "config-data") pod "63366e84-551e-4eb0-a007-7156b6ab1d6b" (UID: "63366e84-551e-4eb0-a007-7156b6ab1d6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.156786 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.156877 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-config-data\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.156904 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.156924 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-run-httpd\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.156959 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9wgw\" (UniqueName: \"kubernetes.io/projected/5e7ade87-ab92-439d-b98c-5cfad50cd547-kube-api-access-w9wgw\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.157016 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-log-httpd\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.157042 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-scripts\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.157102 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.157135 4719 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.157144 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.157153 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.157164 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbj2l\" (UniqueName: \"kubernetes.io/projected/63366e84-551e-4eb0-a007-7156b6ab1d6b-kube-api-access-tbj2l\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.188241 4719 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.235993 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "63366e84-551e-4eb0-a007-7156b6ab1d6b" (UID: "63366e84-551e-4eb0-a007-7156b6ab1d6b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.262794 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-log-httpd\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.262863 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-scripts\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.262891 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.262939 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-config-data\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.262961 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.262981 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-run-httpd\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.263013 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9wgw\" (UniqueName: \"kubernetes.io/projected/5e7ade87-ab92-439d-b98c-5cfad50cd547-kube-api-access-w9wgw\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.263057 4719 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63366e84-551e-4eb0-a007-7156b6ab1d6b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.263069 4719 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.263724 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-log-httpd\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.270522 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-run-httpd\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.276608 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.277535 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-config-data\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.278184 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-scripts\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.291664 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9wgw\" (UniqueName: \"kubernetes.io/projected/5e7ade87-ab92-439d-b98c-5cfad50cd547-kube-api-access-w9wgw\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.294798 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.313132 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"63366e84-551e-4eb0-a007-7156b6ab1d6b","Type":"ContainerDied","Data":"580493cf67a7c6217cf5b3d1230cd2f7eb90d067ae3c56b2533df1dad3da25f3"} Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.313442 4719 scope.go:117] "RemoveContainer" containerID="be535f9264accf5731790a1bfd0b7a52d63b8830d8427c919804c75fedbd7290" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.313669 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.335415 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.351339 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"adf32d03-4559-4d49-8577-c0dac00c1fec","Type":"ContainerDied","Data":"8119c821b57b0b70ed1a77e598c860afdda1bc42a29d05231cdc2739ae30247a"} Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.351442 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.397295 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.420376 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7f6dc9444c-2jcmv" event={"ID":"60352f3a-8ee6-498d-9ee3-152cc4ab562f","Type":"ContainerStarted","Data":"e69947cc65094a1b88a82f045ac64a6bb0455a06df82d38b2d63a2c3a337e16b"} Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.466160 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.499307 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51211ffc-0363-4e48-9972-15b8e997d4de" path="/var/lib/kubelet/pods/51211ffc-0363-4e48-9972-15b8e997d4de/volumes" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.500278 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf7e6ad7-6983-4455-af5d-5f7dda131dd2" path="/var/lib/kubelet/pods/cf7e6ad7-6983-4455-af5d-5f7dda131dd2/volumes" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.501393 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.511925 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.529922 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.534916 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.536491 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.544456 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.568827 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.576308 4719 scope.go:117] "RemoveContainer" containerID="ac4e16b1d34616e8a23c7400e0474377f48d038a2b38a39b0094520b9e4d1a07" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.580396 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.619446 4719 scope.go:117] "RemoveContainer" containerID="03a72fcca616f5e0c4506044788aeee363458eaccd166093db389ab297da0d56" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.644266 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.644319 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.644367 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc2e0fce-d6e0-47e6-bd88-34c45829436a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.644401 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-scripts\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.644428 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb94p\" (UniqueName: \"kubernetes.io/projected/fc2e0fce-d6e0-47e6-bd88-34c45829436a-kube-api-access-nb94p\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.644557 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-config-data\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.644584 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2e0fce-d6e0-47e6-bd88-34c45829436a-logs\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.644711 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.649211 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.662036 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.678623 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.678845 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.678983 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.743022 4719 scope.go:117] "RemoveContainer" containerID="45b06a2e47710e3a4c485c5df45b5a490e7e8faa8b68cc0e55dfd66411a9f377" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.747743 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-config-data\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.747780 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2e0fce-d6e0-47e6-bd88-34c45829436a-logs\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.747816 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-config-data-custom\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.747835 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-config-data\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.747873 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.747891 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.747910 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6133c925-0cd7-489d-8b32-097e9b78ad70-logs\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.747927 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-scripts\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.747964 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znczn\" (UniqueName: \"kubernetes.io/projected/6133c925-0cd7-489d-8b32-097e9b78ad70-kube-api-access-znczn\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.747984 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.748006 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.748024 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc2e0fce-d6e0-47e6-bd88-34c45829436a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.748043 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-scripts\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.748060 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb94p\" (UniqueName: \"kubernetes.io/projected/fc2e0fce-d6e0-47e6-bd88-34c45829436a-kube-api-access-nb94p\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.748095 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6133c925-0cd7-489d-8b32-097e9b78ad70-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.748117 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.748144 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.759968 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc2e0fce-d6e0-47e6-bd88-34c45829436a-logs\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.761778 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.764933 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-config-data\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.765198 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc2e0fce-d6e0-47e6-bd88-34c45829436a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.765235 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.770545 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-scripts\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.771229 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.771923 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc2e0fce-d6e0-47e6-bd88-34c45829436a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.789524 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb94p\" (UniqueName: \"kubernetes.io/projected/fc2e0fce-d6e0-47e6-bd88-34c45829436a-kube-api-access-nb94p\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.809133 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"fc2e0fce-d6e0-47e6-bd88-34c45829436a\") " pod="openstack/glance-default-external-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.849606 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znczn\" (UniqueName: \"kubernetes.io/projected/6133c925-0cd7-489d-8b32-097e9b78ad70-kube-api-access-znczn\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.849715 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6133c925-0cd7-489d-8b32-097e9b78ad70-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.849738 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.849802 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6133c925-0cd7-489d-8b32-097e9b78ad70-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.849847 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.850297 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-config-data-custom\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.850321 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-config-data\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.850345 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.850363 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6133c925-0cd7-489d-8b32-097e9b78ad70-logs\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.850379 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-scripts\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.851702 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6133c925-0cd7-489d-8b32-097e9b78ad70-logs\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.855126 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.858525 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.871264 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-scripts\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.871696 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.875567 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-config-data\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.876237 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znczn\" (UniqueName: \"kubernetes.io/projected/6133c925-0cd7-489d-8b32-097e9b78ad70-kube-api-access-znczn\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.876824 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6133c925-0cd7-489d-8b32-097e9b78ad70-config-data-custom\") pod \"cinder-api-0\" (UID: \"6133c925-0cd7-489d-8b32-097e9b78ad70\") " pod="openstack/cinder-api-0" Dec 15 12:39:33 crc kubenswrapper[4719]: I1215 12:39:33.897607 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.026059 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.204075 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.285298 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.472036 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e7ade87-ab92-439d-b98c-5cfad50cd547","Type":"ContainerStarted","Data":"b3755e3029205ef2a5e12d0378c8aa32659940364c8aa2001304bb7d54f6ad26"} Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.482906 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7f6dc9444c-2jcmv" event={"ID":"60352f3a-8ee6-498d-9ee3-152cc4ab562f","Type":"ContainerStarted","Data":"33e43997d0072db456e1488fcd695aa6fa36778092876e28409ed87346bbb5af"} Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.483707 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.483756 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.504004 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f26a2026-ce1a-4411-8e19-e1981be6f927","Type":"ContainerStarted","Data":"51ab0df69af2d5876d83084a96c234b13f8b84bf845a2f1b9c39ac665747efc7"} Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.524350 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7f6dc9444c-2jcmv" podStartSLOduration=13.524330575 podStartE2EDuration="13.524330575s" podCreationTimestamp="2025-12-15 12:39:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:39:34.511280066 +0000 UTC m=+1335.453573106" watchObservedRunningTime="2025-12-15 12:39:34.524330575 +0000 UTC m=+1335.466623605" Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.613838 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 15 12:39:34 crc kubenswrapper[4719]: W1215 12:39:34.641939 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc2e0fce_d6e0_47e6_bd88_34c45829436a.slice/crio-361d83dc0cfe79f7c2033c48d35de9f90aac255d29d6579fb43cd2ad4e7d6fa0 WatchSource:0}: Error finding container 361d83dc0cfe79f7c2033c48d35de9f90aac255d29d6579fb43cd2ad4e7d6fa0: Status 404 returned error can't find the container with id 361d83dc0cfe79f7c2033c48d35de9f90aac255d29d6579fb43cd2ad4e7d6fa0 Dec 15 12:39:34 crc kubenswrapper[4719]: I1215 12:39:34.656802 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 15 12:39:35 crc kubenswrapper[4719]: I1215 12:39:35.498684 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63366e84-551e-4eb0-a007-7156b6ab1d6b" path="/var/lib/kubelet/pods/63366e84-551e-4eb0-a007-7156b6ab1d6b/volumes" Dec 15 12:39:35 crc kubenswrapper[4719]: I1215 12:39:35.499962 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adf32d03-4559-4d49-8577-c0dac00c1fec" path="/var/lib/kubelet/pods/adf32d03-4559-4d49-8577-c0dac00c1fec/volumes" Dec 15 12:39:35 crc kubenswrapper[4719]: I1215 12:39:35.528139 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6133c925-0cd7-489d-8b32-097e9b78ad70","Type":"ContainerStarted","Data":"ed5313bb1e5f38f669e3fe6a268668f1c701834abf3f5532bbeadb93d8a630a6"} Dec 15 12:39:35 crc kubenswrapper[4719]: I1215 12:39:35.531417 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fc2e0fce-d6e0-47e6-bd88-34c45829436a","Type":"ContainerStarted","Data":"361d83dc0cfe79f7c2033c48d35de9f90aac255d29d6579fb43cd2ad4e7d6fa0"} Dec 15 12:39:36 crc kubenswrapper[4719]: I1215 12:39:36.610371 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fc2e0fce-d6e0-47e6-bd88-34c45829436a","Type":"ContainerStarted","Data":"00f474bb4052b7d45e65c3b1f44fa3e9a02e658ab8e6c86a1ff74bed9798e058"} Dec 15 12:39:36 crc kubenswrapper[4719]: I1215 12:39:36.642326 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e7ade87-ab92-439d-b98c-5cfad50cd547","Type":"ContainerStarted","Data":"7600552b737abcdd16f372c7fe29eada0dce58285cd7bb897810c0ca3e0fd1f7"} Dec 15 12:39:36 crc kubenswrapper[4719]: I1215 12:39:36.652329 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f26a2026-ce1a-4411-8e19-e1981be6f927","Type":"ContainerStarted","Data":"512847b6ed63a71d15a10b7c3824cec0a91214e9386b166edbfa37ad728e43a0"} Dec 15 12:39:36 crc kubenswrapper[4719]: I1215 12:39:36.673030 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6133c925-0cd7-489d-8b32-097e9b78ad70","Type":"ContainerStarted","Data":"a0926c4d52acc34f071f235a0d0058bd5b5545231b3228e78f162f45be0116f1"} Dec 15 12:39:36 crc kubenswrapper[4719]: I1215 12:39:36.721830 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 15 12:39:36 crc kubenswrapper[4719]: I1215 12:39:36.722038 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3887591c-e8a9-43af-b2fb-f2b5cb4e1256" containerName="kube-state-metrics" containerID="cri-o://9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b" gracePeriod=30 Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.390065 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.390548 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.392379 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"e526058563f282b4d65e64a01944ac0f829154bf6a70328411350711beca2af4"} pod="openstack/horizon-7c6bcf58b6-vx99l" containerMessage="Container horizon failed startup probe, will be restarted" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.392438 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" containerID="cri-o://e526058563f282b4d65e64a01944ac0f829154bf6a70328411350711beca2af4" gracePeriod=30 Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.413266 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.568877 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f42lq\" (UniqueName: \"kubernetes.io/projected/3887591c-e8a9-43af-b2fb-f2b5cb4e1256-kube-api-access-f42lq\") pod \"3887591c-e8a9-43af-b2fb-f2b5cb4e1256\" (UID: \"3887591c-e8a9-43af-b2fb-f2b5cb4e1256\") " Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.577974 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc5fb48-jqf6b" podUID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.578131 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3887591c-e8a9-43af-b2fb-f2b5cb4e1256-kube-api-access-f42lq" (OuterVolumeSpecName: "kube-api-access-f42lq") pod "3887591c-e8a9-43af-b2fb-f2b5cb4e1256" (UID: "3887591c-e8a9-43af-b2fb-f2b5cb4e1256"). InnerVolumeSpecName "kube-api-access-f42lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.672163 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f42lq\" (UniqueName: \"kubernetes.io/projected/3887591c-e8a9-43af-b2fb-f2b5cb4e1256-kube-api-access-f42lq\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.682937 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fc2e0fce-d6e0-47e6-bd88-34c45829436a","Type":"ContainerStarted","Data":"13812a4eead8fce66da3535c9152347972ac7702259bcfc2b51206c27cc9c704"} Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.686315 4719 generic.go:334] "Generic (PLEG): container finished" podID="3887591c-e8a9-43af-b2fb-f2b5cb4e1256" containerID="9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b" exitCode=2 Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.686382 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3887591c-e8a9-43af-b2fb-f2b5cb4e1256","Type":"ContainerDied","Data":"9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b"} Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.686410 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3887591c-e8a9-43af-b2fb-f2b5cb4e1256","Type":"ContainerDied","Data":"a4d3dc466d454757d508aeba26d9b613caca6575bc2bbf7109b587dcafed5f8c"} Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.686429 4719 scope.go:117] "RemoveContainer" containerID="9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.686541 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.698121 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e7ade87-ab92-439d-b98c-5cfad50cd547","Type":"ContainerStarted","Data":"3d61693fef24951c548a17a249997c557e284901c963a60aec0c717387eecd61"} Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.716275 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.716256281 podStartE2EDuration="4.716256281s" podCreationTimestamp="2025-12-15 12:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:39:37.711367748 +0000 UTC m=+1338.653660788" watchObservedRunningTime="2025-12-15 12:39:37.716256281 +0000 UTC m=+1338.658549311" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.740175 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f26a2026-ce1a-4411-8e19-e1981be6f927","Type":"ContainerStarted","Data":"5523ccb3235cbf8f330037fd08df4197556b8a032e38b344028f77de49028605"} Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.764929 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.779065 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6133c925-0cd7-489d-8b32-097e9b78ad70","Type":"ContainerStarted","Data":"ef9f04896c6c5652ed94f32b01594b07ea14b742c17c64e36e470a9f7a640eec"} Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.779654 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.789268 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.808215 4719 scope.go:117] "RemoveContainer" containerID="9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b" Dec 15 12:39:37 crc kubenswrapper[4719]: E1215 12:39:37.808551 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b\": container with ID starting with 9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b not found: ID does not exist" containerID="9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.808590 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b"} err="failed to get container status \"9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b\": rpc error: code = NotFound desc = could not find container \"9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b\": container with ID starting with 9871f7683cab08d5cfeb27ad68877a6c46a9f210853937bf74d5c4770b27e07b not found: ID does not exist" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.824438 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 15 12:39:37 crc kubenswrapper[4719]: E1215 12:39:37.824913 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3887591c-e8a9-43af-b2fb-f2b5cb4e1256" containerName="kube-state-metrics" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.824931 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="3887591c-e8a9-43af-b2fb-f2b5cb4e1256" containerName="kube-state-metrics" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.825116 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="3887591c-e8a9-43af-b2fb-f2b5cb4e1256" containerName="kube-state-metrics" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.825716 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.831729 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.832491 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.835205 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.838941 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.838923196 podStartE2EDuration="5.838923196s" podCreationTimestamp="2025-12-15 12:39:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:39:37.766766814 +0000 UTC m=+1338.709059844" watchObservedRunningTime="2025-12-15 12:39:37.838923196 +0000 UTC m=+1338.781216226" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.849029 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.849013023 podStartE2EDuration="4.849013023s" podCreationTimestamp="2025-12-15 12:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:39:37.802790243 +0000 UTC m=+1338.745083273" watchObservedRunningTime="2025-12-15 12:39:37.849013023 +0000 UTC m=+1338.791306053" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.979390 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9a236191-195f-4941-b740-ac7ee096858a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.979461 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a236191-195f-4941-b740-ac7ee096858a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.979542 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a236191-195f-4941-b740-ac7ee096858a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:37 crc kubenswrapper[4719]: I1215 12:39:37.979561 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vbc5\" (UniqueName: \"kubernetes.io/projected/9a236191-195f-4941-b740-ac7ee096858a-kube-api-access-4vbc5\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.081138 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9a236191-195f-4941-b740-ac7ee096858a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.081208 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a236191-195f-4941-b740-ac7ee096858a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.081286 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a236191-195f-4941-b740-ac7ee096858a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.081305 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vbc5\" (UniqueName: \"kubernetes.io/projected/9a236191-195f-4941-b740-ac7ee096858a-kube-api-access-4vbc5\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.086458 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9a236191-195f-4941-b740-ac7ee096858a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.087031 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a236191-195f-4941-b740-ac7ee096858a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.087522 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a236191-195f-4941-b740-ac7ee096858a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.102746 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vbc5\" (UniqueName: \"kubernetes.io/projected/9a236191-195f-4941-b740-ac7ee096858a-kube-api-access-4vbc5\") pod \"kube-state-metrics-0\" (UID: \"9a236191-195f-4941-b740-ac7ee096858a\") " pod="openstack/kube-state-metrics-0" Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.162624 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.727643 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 15 12:39:38 crc kubenswrapper[4719]: W1215 12:39:38.741053 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a236191_195f_4941_b740_ac7ee096858a.slice/crio-25ff4c9188fe895c324bd6058f08967c29ec4ca594cc87673a151c065bc6b68c WatchSource:0}: Error finding container 25ff4c9188fe895c324bd6058f08967c29ec4ca594cc87673a151c065bc6b68c: Status 404 returned error can't find the container with id 25ff4c9188fe895c324bd6058f08967c29ec4ca594cc87673a151c065bc6b68c Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.812056 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e7ade87-ab92-439d-b98c-5cfad50cd547","Type":"ContainerStarted","Data":"9de7bf493ed1aeea89bec6454411419ec533478d5c7a000439758c78139947c5"} Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.819092 4719 generic.go:334] "Generic (PLEG): container finished" podID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" containerID="6c8f61185a298bc47415e2490728b2af74efe149d868a8e1c30bbe335d4b351b" exitCode=0 Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.819163 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5649967964-ctb22" event={"ID":"2a1fc07a-9062-4ed7-8610-065fa84f1b2b","Type":"ContainerDied","Data":"6c8f61185a298bc47415e2490728b2af74efe149d868a8e1c30bbe335d4b351b"} Dec 15 12:39:38 crc kubenswrapper[4719]: I1215 12:39:38.820931 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9a236191-195f-4941-b740-ac7ee096858a","Type":"ContainerStarted","Data":"25ff4c9188fe895c324bd6058f08967c29ec4ca594cc87673a151c065bc6b68c"} Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.070964 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5649967964-ctb22" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.148467 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-config\") pod \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.148533 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-httpd-config\") pod \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.148631 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtrjk\" (UniqueName: \"kubernetes.io/projected/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-kube-api-access-wtrjk\") pod \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.148711 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-combined-ca-bundle\") pod \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.148739 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-ovndb-tls-certs\") pod \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\" (UID: \"2a1fc07a-9062-4ed7-8610-065fa84f1b2b\") " Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.161259 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-kube-api-access-wtrjk" (OuterVolumeSpecName: "kube-api-access-wtrjk") pod "2a1fc07a-9062-4ed7-8610-065fa84f1b2b" (UID: "2a1fc07a-9062-4ed7-8610-065fa84f1b2b"). InnerVolumeSpecName "kube-api-access-wtrjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.162996 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "2a1fc07a-9062-4ed7-8610-065fa84f1b2b" (UID: "2a1fc07a-9062-4ed7-8610-065fa84f1b2b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.255473 4719 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.255505 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtrjk\" (UniqueName: \"kubernetes.io/projected/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-kube-api-access-wtrjk\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.300018 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a1fc07a-9062-4ed7-8610-065fa84f1b2b" (UID: "2a1fc07a-9062-4ed7-8610-065fa84f1b2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.330095 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-config" (OuterVolumeSpecName: "config") pod "2a1fc07a-9062-4ed7-8610-065fa84f1b2b" (UID: "2a1fc07a-9062-4ed7-8610-065fa84f1b2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.358694 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.358738 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.365818 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "2a1fc07a-9062-4ed7-8610-065fa84f1b2b" (UID: "2a1fc07a-9062-4ed7-8610-065fa84f1b2b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.460753 4719 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a1fc07a-9062-4ed7-8610-065fa84f1b2b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.489166 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3887591c-e8a9-43af-b2fb-f2b5cb4e1256" path="/var/lib/kubelet/pods/3887591c-e8a9-43af-b2fb-f2b5cb4e1256/volumes" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.834914 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9a236191-195f-4941-b740-ac7ee096858a","Type":"ContainerStarted","Data":"718833a668fa3f06b00d8672bf66beb765b00d37e6a92f84eb3f60c452373bdb"} Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.835957 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.845661 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5649967964-ctb22" event={"ID":"2a1fc07a-9062-4ed7-8610-065fa84f1b2b","Type":"ContainerDied","Data":"b7f4cef3a00f011ddf87b7a272684ad089ce3cba6f665e4deaf96c24edb74ae9"} Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.845710 4719 scope.go:117] "RemoveContainer" containerID="594ed6de5e5587a3240d090ec7dcad0447ce09468e5b92aef60bbc735df86854" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.845893 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5649967964-ctb22" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.857771 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.474717263 podStartE2EDuration="2.857755769s" podCreationTimestamp="2025-12-15 12:39:37 +0000 UTC" firstStartedPulling="2025-12-15 12:39:38.743383268 +0000 UTC m=+1339.685676288" lastFinishedPulling="2025-12-15 12:39:39.126421764 +0000 UTC m=+1340.068714794" observedRunningTime="2025-12-15 12:39:39.856239792 +0000 UTC m=+1340.798532822" watchObservedRunningTime="2025-12-15 12:39:39.857755769 +0000 UTC m=+1340.800048799" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.885602 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5649967964-ctb22"] Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.888923 4719 scope.go:117] "RemoveContainer" containerID="6c8f61185a298bc47415e2490728b2af74efe149d868a8e1c30bbe335d4b351b" Dec 15 12:39:39 crc kubenswrapper[4719]: I1215 12:39:39.899606 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5649967964-ctb22"] Dec 15 12:39:40 crc kubenswrapper[4719]: I1215 12:39:40.214069 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:41 crc kubenswrapper[4719]: I1215 12:39:41.489633 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" path="/var/lib/kubelet/pods/2a1fc07a-9062-4ed7-8610-065fa84f1b2b/volumes" Dec 15 12:39:41 crc kubenswrapper[4719]: I1215 12:39:41.592728 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:41 crc kubenswrapper[4719]: I1215 12:39:41.593554 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7f6dc9444c-2jcmv" Dec 15 12:39:41 crc kubenswrapper[4719]: I1215 12:39:41.864069 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e7ade87-ab92-439d-b98c-5cfad50cd547","Type":"ContainerStarted","Data":"707b6808f28eb867644ae3eda35b23aa8d0fc3d058511201db9ffd190cd497d0"} Dec 15 12:39:41 crc kubenswrapper[4719]: I1215 12:39:41.864260 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="proxy-httpd" containerID="cri-o://707b6808f28eb867644ae3eda35b23aa8d0fc3d058511201db9ffd190cd497d0" gracePeriod=30 Dec 15 12:39:41 crc kubenswrapper[4719]: I1215 12:39:41.864260 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="ceilometer-central-agent" containerID="cri-o://7600552b737abcdd16f372c7fe29eada0dce58285cd7bb897810c0ca3e0fd1f7" gracePeriod=30 Dec 15 12:39:41 crc kubenswrapper[4719]: I1215 12:39:41.864324 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="sg-core" containerID="cri-o://9de7bf493ed1aeea89bec6454411419ec533478d5c7a000439758c78139947c5" gracePeriod=30 Dec 15 12:39:41 crc kubenswrapper[4719]: I1215 12:39:41.864394 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="ceilometer-notification-agent" containerID="cri-o://3d61693fef24951c548a17a249997c557e284901c963a60aec0c717387eecd61" gracePeriod=30 Dec 15 12:39:41 crc kubenswrapper[4719]: I1215 12:39:41.884469 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.221156978 podStartE2EDuration="9.884408928s" podCreationTimestamp="2025-12-15 12:39:32 +0000 UTC" firstStartedPulling="2025-12-15 12:39:34.171107852 +0000 UTC m=+1335.113400882" lastFinishedPulling="2025-12-15 12:39:40.834359802 +0000 UTC m=+1341.776652832" observedRunningTime="2025-12-15 12:39:41.88222179 +0000 UTC m=+1342.824514840" watchObservedRunningTime="2025-12-15 12:39:41.884408928 +0000 UTC m=+1342.826701958" Dec 15 12:39:42 crc kubenswrapper[4719]: I1215 12:39:42.895344 4719 generic.go:334] "Generic (PLEG): container finished" podID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerID="707b6808f28eb867644ae3eda35b23aa8d0fc3d058511201db9ffd190cd497d0" exitCode=0 Dec 15 12:39:42 crc kubenswrapper[4719]: I1215 12:39:42.895381 4719 generic.go:334] "Generic (PLEG): container finished" podID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerID="9de7bf493ed1aeea89bec6454411419ec533478d5c7a000439758c78139947c5" exitCode=2 Dec 15 12:39:42 crc kubenswrapper[4719]: I1215 12:39:42.895389 4719 generic.go:334] "Generic (PLEG): container finished" podID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerID="3d61693fef24951c548a17a249997c557e284901c963a60aec0c717387eecd61" exitCode=0 Dec 15 12:39:42 crc kubenswrapper[4719]: I1215 12:39:42.895419 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e7ade87-ab92-439d-b98c-5cfad50cd547","Type":"ContainerDied","Data":"707b6808f28eb867644ae3eda35b23aa8d0fc3d058511201db9ffd190cd497d0"} Dec 15 12:39:42 crc kubenswrapper[4719]: I1215 12:39:42.895459 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e7ade87-ab92-439d-b98c-5cfad50cd547","Type":"ContainerDied","Data":"9de7bf493ed1aeea89bec6454411419ec533478d5c7a000439758c78139947c5"} Dec 15 12:39:42 crc kubenswrapper[4719]: I1215 12:39:42.895473 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e7ade87-ab92-439d-b98c-5cfad50cd547","Type":"ContainerDied","Data":"3d61693fef24951c548a17a249997c557e284901c963a60aec0c717387eecd61"} Dec 15 12:39:43 crc kubenswrapper[4719]: I1215 12:39:43.398086 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:43 crc kubenswrapper[4719]: I1215 12:39:43.398433 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:43 crc kubenswrapper[4719]: I1215 12:39:43.429767 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:43 crc kubenswrapper[4719]: I1215 12:39:43.447464 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:43 crc kubenswrapper[4719]: I1215 12:39:43.898201 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 15 12:39:43 crc kubenswrapper[4719]: I1215 12:39:43.898781 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 15 12:39:43 crc kubenswrapper[4719]: I1215 12:39:43.901978 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:43 crc kubenswrapper[4719]: I1215 12:39:43.902026 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:43 crc kubenswrapper[4719]: I1215 12:39:43.936576 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 15 12:39:43 crc kubenswrapper[4719]: I1215 12:39:43.950695 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 15 12:39:44 crc kubenswrapper[4719]: I1215 12:39:44.944408 4719 generic.go:334] "Generic (PLEG): container finished" podID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerID="7600552b737abcdd16f372c7fe29eada0dce58285cd7bb897810c0ca3e0fd1f7" exitCode=0 Dec 15 12:39:44 crc kubenswrapper[4719]: I1215 12:39:44.945625 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e7ade87-ab92-439d-b98c-5cfad50cd547","Type":"ContainerDied","Data":"7600552b737abcdd16f372c7fe29eada0dce58285cd7bb897810c0ca3e0fd1f7"} Dec 15 12:39:44 crc kubenswrapper[4719]: I1215 12:39:44.946356 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 15 12:39:44 crc kubenswrapper[4719]: I1215 12:39:44.946606 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.115361 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.170313 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9wgw\" (UniqueName: \"kubernetes.io/projected/5e7ade87-ab92-439d-b98c-5cfad50cd547-kube-api-access-w9wgw\") pod \"5e7ade87-ab92-439d-b98c-5cfad50cd547\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.170373 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-sg-core-conf-yaml\") pod \"5e7ade87-ab92-439d-b98c-5cfad50cd547\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.170444 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-combined-ca-bundle\") pod \"5e7ade87-ab92-439d-b98c-5cfad50cd547\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.170522 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-run-httpd\") pod \"5e7ade87-ab92-439d-b98c-5cfad50cd547\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.170616 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-config-data\") pod \"5e7ade87-ab92-439d-b98c-5cfad50cd547\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.170756 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-log-httpd\") pod \"5e7ade87-ab92-439d-b98c-5cfad50cd547\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.170785 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-scripts\") pod \"5e7ade87-ab92-439d-b98c-5cfad50cd547\" (UID: \"5e7ade87-ab92-439d-b98c-5cfad50cd547\") " Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.176832 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5e7ade87-ab92-439d-b98c-5cfad50cd547" (UID: "5e7ade87-ab92-439d-b98c-5cfad50cd547"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.177169 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5e7ade87-ab92-439d-b98c-5cfad50cd547" (UID: "5e7ade87-ab92-439d-b98c-5cfad50cd547"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.194015 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-scripts" (OuterVolumeSpecName: "scripts") pod "5e7ade87-ab92-439d-b98c-5cfad50cd547" (UID: "5e7ade87-ab92-439d-b98c-5cfad50cd547"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.194115 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e7ade87-ab92-439d-b98c-5cfad50cd547-kube-api-access-w9wgw" (OuterVolumeSpecName: "kube-api-access-w9wgw") pod "5e7ade87-ab92-439d-b98c-5cfad50cd547" (UID: "5e7ade87-ab92-439d-b98c-5cfad50cd547"). InnerVolumeSpecName "kube-api-access-w9wgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.261948 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5e7ade87-ab92-439d-b98c-5cfad50cd547" (UID: "5e7ade87-ab92-439d-b98c-5cfad50cd547"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.275223 4719 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.275256 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.275265 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9wgw\" (UniqueName: \"kubernetes.io/projected/5e7ade87-ab92-439d-b98c-5cfad50cd547-kube-api-access-w9wgw\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.275277 4719 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.275285 4719 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5e7ade87-ab92-439d-b98c-5cfad50cd547-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.385414 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-config-data" (OuterVolumeSpecName: "config-data") pod "5e7ade87-ab92-439d-b98c-5cfad50cd547" (UID: "5e7ade87-ab92-439d-b98c-5cfad50cd547"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.393021 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e7ade87-ab92-439d-b98c-5cfad50cd547" (UID: "5e7ade87-ab92-439d-b98c-5cfad50cd547"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.481568 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.481606 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e7ade87-ab92-439d-b98c-5cfad50cd547-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.957326 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.958224 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5e7ade87-ab92-439d-b98c-5cfad50cd547","Type":"ContainerDied","Data":"b3755e3029205ef2a5e12d0378c8aa32659940364c8aa2001304bb7d54f6ad26"} Dec 15 12:39:45 crc kubenswrapper[4719]: I1215 12:39:45.958271 4719 scope.go:117] "RemoveContainer" containerID="707b6808f28eb867644ae3eda35b23aa8d0fc3d058511201db9ffd190cd497d0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.010775 4719 scope.go:117] "RemoveContainer" containerID="9de7bf493ed1aeea89bec6454411419ec533478d5c7a000439758c78139947c5" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.057586 4719 scope.go:117] "RemoveContainer" containerID="3d61693fef24951c548a17a249997c557e284901c963a60aec0c717387eecd61" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.086420 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.106414 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.133513 4719 scope.go:117] "RemoveContainer" containerID="7600552b737abcdd16f372c7fe29eada0dce58285cd7bb897810c0ca3e0fd1f7" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.133640 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:46 crc kubenswrapper[4719]: E1215 12:39:46.133981 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="sg-core" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.133992 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="sg-core" Dec 15 12:39:46 crc kubenswrapper[4719]: E1215 12:39:46.134002 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" containerName="neutron-httpd" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134007 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" containerName="neutron-httpd" Dec 15 12:39:46 crc kubenswrapper[4719]: E1215 12:39:46.134034 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="proxy-httpd" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134040 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="proxy-httpd" Dec 15 12:39:46 crc kubenswrapper[4719]: E1215 12:39:46.134054 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="ceilometer-notification-agent" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134059 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="ceilometer-notification-agent" Dec 15 12:39:46 crc kubenswrapper[4719]: E1215 12:39:46.134071 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" containerName="neutron-api" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134076 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" containerName="neutron-api" Dec 15 12:39:46 crc kubenswrapper[4719]: E1215 12:39:46.134087 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="ceilometer-central-agent" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134092 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="ceilometer-central-agent" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134247 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" containerName="neutron-httpd" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134265 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a1fc07a-9062-4ed7-8610-065fa84f1b2b" containerName="neutron-api" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134273 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="proxy-httpd" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134283 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="sg-core" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134295 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="ceilometer-notification-agent" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.134304 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" containerName="ceilometer-central-agent" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.135772 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.138590 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.139908 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.140177 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.140411 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.203306 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.203558 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-run-httpd\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.203649 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-log-httpd\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.203735 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.203809 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-config-data\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.203921 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzcbn\" (UniqueName: \"kubernetes.io/projected/c6dabdcb-160d-47b8-b616-bc13880955da-kube-api-access-qzcbn\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.204016 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-scripts\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.204100 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.305479 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.305532 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.305554 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-run-httpd\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.305589 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-log-httpd\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.305621 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.305637 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-config-data\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.305684 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzcbn\" (UniqueName: \"kubernetes.io/projected/c6dabdcb-160d-47b8-b616-bc13880955da-kube-api-access-qzcbn\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.305727 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-scripts\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.307052 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-run-httpd\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.307286 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-log-httpd\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.312261 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-scripts\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.313497 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-config-data\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.314002 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.314572 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.335449 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.350500 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzcbn\" (UniqueName: \"kubernetes.io/projected/c6dabdcb-160d-47b8-b616-bc13880955da-kube-api-access-qzcbn\") pod \"ceilometer-0\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.465314 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.482234 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.812516 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 15 12:39:46 crc kubenswrapper[4719]: I1215 12:39:46.988767 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:47 crc kubenswrapper[4719]: I1215 12:39:47.485486 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e7ade87-ab92-439d-b98c-5cfad50cd547" path="/var/lib/kubelet/pods/5e7ade87-ab92-439d-b98c-5cfad50cd547/volumes" Dec 15 12:39:47 crc kubenswrapper[4719]: I1215 12:39:47.975810 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6dabdcb-160d-47b8-b616-bc13880955da","Type":"ContainerStarted","Data":"da539c211ccd4b0ca0161de7dcb8b6eb60f2db17bbdeca541d9d253856e352fb"} Dec 15 12:39:48 crc kubenswrapper[4719]: I1215 12:39:48.260015 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 15 12:39:48 crc kubenswrapper[4719]: I1215 12:39:48.794515 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:48 crc kubenswrapper[4719]: I1215 12:39:48.794623 4719 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 15 12:39:49 crc kubenswrapper[4719]: I1215 12:39:49.097852 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 15 12:39:49 crc kubenswrapper[4719]: I1215 12:39:49.672435 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 15 12:39:49 crc kubenswrapper[4719]: I1215 12:39:49.672551 4719 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 15 12:39:49 crc kubenswrapper[4719]: I1215 12:39:49.727408 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 15 12:39:51 crc kubenswrapper[4719]: I1215 12:39:51.357013 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:39:51 crc kubenswrapper[4719]: I1215 12:39:51.357543 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:39:51 crc kubenswrapper[4719]: I1215 12:39:51.357588 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:39:51 crc kubenswrapper[4719]: I1215 12:39:51.358264 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e82d238054920207f8e0cb0360298c4a0049bf10cc2919ab754c599ae7031178"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 12:39:51 crc kubenswrapper[4719]: I1215 12:39:51.358315 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://e82d238054920207f8e0cb0360298c4a0049bf10cc2919ab754c599ae7031178" gracePeriod=600 Dec 15 12:39:52 crc kubenswrapper[4719]: I1215 12:39:52.029810 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6dabdcb-160d-47b8-b616-bc13880955da","Type":"ContainerStarted","Data":"705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543"} Dec 15 12:39:52 crc kubenswrapper[4719]: I1215 12:39:52.031970 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0b238728-d6aa-4b27-b48b-beb0cc7ff437","Type":"ContainerStarted","Data":"3cba2e24bf28faaeb6fdcd0eddb366f2c5203e45aee3fceec9141febd835a3aa"} Dec 15 12:39:52 crc kubenswrapper[4719]: I1215 12:39:52.036134 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="e82d238054920207f8e0cb0360298c4a0049bf10cc2919ab754c599ae7031178" exitCode=0 Dec 15 12:39:52 crc kubenswrapper[4719]: I1215 12:39:52.036169 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"e82d238054920207f8e0cb0360298c4a0049bf10cc2919ab754c599ae7031178"} Dec 15 12:39:52 crc kubenswrapper[4719]: I1215 12:39:52.036195 4719 scope.go:117] "RemoveContainer" containerID="6512c03468894d57c0ec7081d538364858ba849e9d66b22fa695a25587406622" Dec 15 12:39:52 crc kubenswrapper[4719]: I1215 12:39:52.054737 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.201410756 podStartE2EDuration="40.054720893s" podCreationTimestamp="2025-12-15 12:39:12 +0000 UTC" firstStartedPulling="2025-12-15 12:39:13.725951736 +0000 UTC m=+1314.668244766" lastFinishedPulling="2025-12-15 12:39:50.579261863 +0000 UTC m=+1351.521554903" observedRunningTime="2025-12-15 12:39:52.048375164 +0000 UTC m=+1352.990668194" watchObservedRunningTime="2025-12-15 12:39:52.054720893 +0000 UTC m=+1352.997013923" Dec 15 12:39:52 crc kubenswrapper[4719]: I1215 12:39:52.620664 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc5fb48-jqf6b" podUID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:39:52 crc kubenswrapper[4719]: I1215 12:39:52.620965 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:39:52 crc kubenswrapper[4719]: I1215 12:39:52.621910 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"b3d3c0d45c63097362f15fdd3a23877d00f86c6a1a144fb9d5b2f5b8da10fbb1"} pod="openstack/horizon-6bc5fb48-jqf6b" containerMessage="Container horizon failed startup probe, will be restarted" Dec 15 12:39:52 crc kubenswrapper[4719]: I1215 12:39:52.621946 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6bc5fb48-jqf6b" podUID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerName="horizon" containerID="cri-o://b3d3c0d45c63097362f15fdd3a23877d00f86c6a1a144fb9d5b2f5b8da10fbb1" gracePeriod=30 Dec 15 12:39:53 crc kubenswrapper[4719]: I1215 12:39:53.046882 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b"} Dec 15 12:39:54 crc kubenswrapper[4719]: I1215 12:39:54.056328 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6dabdcb-160d-47b8-b616-bc13880955da","Type":"ContainerStarted","Data":"6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30"} Dec 15 12:39:54 crc kubenswrapper[4719]: I1215 12:39:54.056953 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6dabdcb-160d-47b8-b616-bc13880955da","Type":"ContainerStarted","Data":"8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb"} Dec 15 12:39:55 crc kubenswrapper[4719]: I1215 12:39:55.274148 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:39:56 crc kubenswrapper[4719]: I1215 12:39:56.077808 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6dabdcb-160d-47b8-b616-bc13880955da","Type":"ContainerStarted","Data":"ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a"} Dec 15 12:39:56 crc kubenswrapper[4719]: I1215 12:39:56.078003 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="ceilometer-central-agent" containerID="cri-o://705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543" gracePeriod=30 Dec 15 12:39:56 crc kubenswrapper[4719]: I1215 12:39:56.078248 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="proxy-httpd" containerID="cri-o://ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a" gracePeriod=30 Dec 15 12:39:56 crc kubenswrapper[4719]: I1215 12:39:56.078303 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="ceilometer-notification-agent" containerID="cri-o://8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb" gracePeriod=30 Dec 15 12:39:56 crc kubenswrapper[4719]: I1215 12:39:56.078320 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="sg-core" containerID="cri-o://6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30" gracePeriod=30 Dec 15 12:39:56 crc kubenswrapper[4719]: I1215 12:39:56.087100 4719 generic.go:334] "Generic (PLEG): container finished" podID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerID="b3d3c0d45c63097362f15fdd3a23877d00f86c6a1a144fb9d5b2f5b8da10fbb1" exitCode=0 Dec 15 12:39:56 crc kubenswrapper[4719]: I1215 12:39:56.087150 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc5fb48-jqf6b" event={"ID":"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926","Type":"ContainerDied","Data":"b3d3c0d45c63097362f15fdd3a23877d00f86c6a1a144fb9d5b2f5b8da10fbb1"} Dec 15 12:39:56 crc kubenswrapper[4719]: I1215 12:39:56.087192 4719 scope.go:117] "RemoveContainer" containerID="975b276955c100d7fd0352b04978b9b7d4a57b2a3ddba6e0cbb4fd095434fee3" Dec 15 12:39:56 crc kubenswrapper[4719]: I1215 12:39:56.109918 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.836230877 podStartE2EDuration="10.109893609s" podCreationTimestamp="2025-12-15 12:39:46 +0000 UTC" firstStartedPulling="2025-12-15 12:39:47.00879412 +0000 UTC m=+1347.951087150" lastFinishedPulling="2025-12-15 12:39:55.282456852 +0000 UTC m=+1356.224749882" observedRunningTime="2025-12-15 12:39:56.105912914 +0000 UTC m=+1357.048205944" watchObservedRunningTime="2025-12-15 12:39:56.109893609 +0000 UTC m=+1357.052186649" Dec 15 12:39:57 crc kubenswrapper[4719]: I1215 12:39:57.097924 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bc5fb48-jqf6b" event={"ID":"f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926","Type":"ContainerStarted","Data":"d333b807632b15272858efba6e5d917d848dd052737bfa8e3ac75da76e81e1e3"} Dec 15 12:39:57 crc kubenswrapper[4719]: I1215 12:39:57.100637 4719 generic.go:334] "Generic (PLEG): container finished" podID="c6dabdcb-160d-47b8-b616-bc13880955da" containerID="ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a" exitCode=0 Dec 15 12:39:57 crc kubenswrapper[4719]: I1215 12:39:57.100667 4719 generic.go:334] "Generic (PLEG): container finished" podID="c6dabdcb-160d-47b8-b616-bc13880955da" containerID="6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30" exitCode=2 Dec 15 12:39:57 crc kubenswrapper[4719]: I1215 12:39:57.100676 4719 generic.go:334] "Generic (PLEG): container finished" podID="c6dabdcb-160d-47b8-b616-bc13880955da" containerID="8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb" exitCode=0 Dec 15 12:39:57 crc kubenswrapper[4719]: I1215 12:39:57.100696 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6dabdcb-160d-47b8-b616-bc13880955da","Type":"ContainerDied","Data":"ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a"} Dec 15 12:39:57 crc kubenswrapper[4719]: I1215 12:39:57.100721 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6dabdcb-160d-47b8-b616-bc13880955da","Type":"ContainerDied","Data":"6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30"} Dec 15 12:39:57 crc kubenswrapper[4719]: I1215 12:39:57.100731 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6dabdcb-160d-47b8-b616-bc13880955da","Type":"ContainerDied","Data":"8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb"} Dec 15 12:39:57 crc kubenswrapper[4719]: I1215 12:39:57.574843 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:39:57 crc kubenswrapper[4719]: I1215 12:39:57.575161 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.024018 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.164301 4719 generic.go:334] "Generic (PLEG): container finished" podID="c6dabdcb-160d-47b8-b616-bc13880955da" containerID="705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543" exitCode=0 Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.164348 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6dabdcb-160d-47b8-b616-bc13880955da","Type":"ContainerDied","Data":"705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543"} Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.164397 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c6dabdcb-160d-47b8-b616-bc13880955da","Type":"ContainerDied","Data":"da539c211ccd4b0ca0161de7dcb8b6eb60f2db17bbdeca541d9d253856e352fb"} Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.164415 4719 scope.go:117] "RemoveContainer" containerID="ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.164605 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.177840 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-log-httpd\") pod \"c6dabdcb-160d-47b8-b616-bc13880955da\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.178133 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-run-httpd\") pod \"c6dabdcb-160d-47b8-b616-bc13880955da\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.178299 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-combined-ca-bundle\") pod \"c6dabdcb-160d-47b8-b616-bc13880955da\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.178464 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-sg-core-conf-yaml\") pod \"c6dabdcb-160d-47b8-b616-bc13880955da\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.178622 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-ceilometer-tls-certs\") pod \"c6dabdcb-160d-47b8-b616-bc13880955da\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.178829 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-scripts\") pod \"c6dabdcb-160d-47b8-b616-bc13880955da\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.179168 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzcbn\" (UniqueName: \"kubernetes.io/projected/c6dabdcb-160d-47b8-b616-bc13880955da-kube-api-access-qzcbn\") pod \"c6dabdcb-160d-47b8-b616-bc13880955da\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.179487 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-config-data\") pod \"c6dabdcb-160d-47b8-b616-bc13880955da\" (UID: \"c6dabdcb-160d-47b8-b616-bc13880955da\") " Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.186999 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-scripts" (OuterVolumeSpecName: "scripts") pod "c6dabdcb-160d-47b8-b616-bc13880955da" (UID: "c6dabdcb-160d-47b8-b616-bc13880955da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.190896 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c6dabdcb-160d-47b8-b616-bc13880955da" (UID: "c6dabdcb-160d-47b8-b616-bc13880955da"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.191431 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c6dabdcb-160d-47b8-b616-bc13880955da" (UID: "c6dabdcb-160d-47b8-b616-bc13880955da"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.191591 4719 scope.go:117] "RemoveContainer" containerID="6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.206051 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6dabdcb-160d-47b8-b616-bc13880955da-kube-api-access-qzcbn" (OuterVolumeSpecName: "kube-api-access-qzcbn") pod "c6dabdcb-160d-47b8-b616-bc13880955da" (UID: "c6dabdcb-160d-47b8-b616-bc13880955da"). InnerVolumeSpecName "kube-api-access-qzcbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.255940 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c6dabdcb-160d-47b8-b616-bc13880955da" (UID: "c6dabdcb-160d-47b8-b616-bc13880955da"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.256975 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c6dabdcb-160d-47b8-b616-bc13880955da" (UID: "c6dabdcb-160d-47b8-b616-bc13880955da"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.283359 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzcbn\" (UniqueName: \"kubernetes.io/projected/c6dabdcb-160d-47b8-b616-bc13880955da-kube-api-access-qzcbn\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.283393 4719 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.283404 4719 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6dabdcb-160d-47b8-b616-bc13880955da-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.283412 4719 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.283420 4719 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.283429 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.284365 4719 scope.go:117] "RemoveContainer" containerID="8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.287021 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6dabdcb-160d-47b8-b616-bc13880955da" (UID: "c6dabdcb-160d-47b8-b616-bc13880955da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.311225 4719 scope.go:117] "RemoveContainer" containerID="705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.316712 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-config-data" (OuterVolumeSpecName: "config-data") pod "c6dabdcb-160d-47b8-b616-bc13880955da" (UID: "c6dabdcb-160d-47b8-b616-bc13880955da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.336038 4719 scope.go:117] "RemoveContainer" containerID="ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a" Dec 15 12:40:05 crc kubenswrapper[4719]: E1215 12:40:05.336744 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a\": container with ID starting with ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a not found: ID does not exist" containerID="ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.336792 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a"} err="failed to get container status \"ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a\": rpc error: code = NotFound desc = could not find container \"ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a\": container with ID starting with ae5230d32480a1b87e9f7aa659ec7b8b942bcd054b5b2e0352522b0b072b273a not found: ID does not exist" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.336812 4719 scope.go:117] "RemoveContainer" containerID="6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30" Dec 15 12:40:05 crc kubenswrapper[4719]: E1215 12:40:05.337289 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30\": container with ID starting with 6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30 not found: ID does not exist" containerID="6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.337334 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30"} err="failed to get container status \"6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30\": rpc error: code = NotFound desc = could not find container \"6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30\": container with ID starting with 6f3f37538081b840072d7c925d33f21951de57fda3493d59d896c6cb26058e30 not found: ID does not exist" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.337349 4719 scope.go:117] "RemoveContainer" containerID="8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb" Dec 15 12:40:05 crc kubenswrapper[4719]: E1215 12:40:05.337837 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb\": container with ID starting with 8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb not found: ID does not exist" containerID="8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.337882 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb"} err="failed to get container status \"8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb\": rpc error: code = NotFound desc = could not find container \"8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb\": container with ID starting with 8e9091f1c5f69618f840c4e25eff8c5c8d6add8c761bd19cc291e9b82bfb6ceb not found: ID does not exist" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.337895 4719 scope.go:117] "RemoveContainer" containerID="705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543" Dec 15 12:40:05 crc kubenswrapper[4719]: E1215 12:40:05.338225 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543\": container with ID starting with 705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543 not found: ID does not exist" containerID="705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.338263 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543"} err="failed to get container status \"705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543\": rpc error: code = NotFound desc = could not find container \"705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543\": container with ID starting with 705ea0d819bc8a8741e7295db4b6e17211a91c757e77a48b7cb4aff50cd5e543 not found: ID does not exist" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.384697 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.384970 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6dabdcb-160d-47b8-b616-bc13880955da-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.516301 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.531216 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.544633 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:05 crc kubenswrapper[4719]: E1215 12:40:05.545036 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="proxy-httpd" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.545051 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="proxy-httpd" Dec 15 12:40:05 crc kubenswrapper[4719]: E1215 12:40:05.545067 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="ceilometer-central-agent" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.545073 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="ceilometer-central-agent" Dec 15 12:40:05 crc kubenswrapper[4719]: E1215 12:40:05.545084 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="sg-core" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.545090 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="sg-core" Dec 15 12:40:05 crc kubenswrapper[4719]: E1215 12:40:05.545101 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="ceilometer-notification-agent" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.545106 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="ceilometer-notification-agent" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.545275 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="proxy-httpd" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.545289 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="ceilometer-notification-agent" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.545307 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="sg-core" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.545316 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" containerName="ceilometer-central-agent" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.546747 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.551718 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.551975 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.554723 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.582112 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.690624 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.690696 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-config-data\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.690792 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-scripts\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.690879 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.690918 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-log-httpd\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.690936 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.690980 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dw6r\" (UniqueName: \"kubernetes.io/projected/0eb4fa70-5bc2-4349-871f-1c095c1981ab-kube-api-access-8dw6r\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.691051 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-run-httpd\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.793192 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-config-data\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.793239 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-scripts\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.793262 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.793283 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-log-httpd\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.793307 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.793332 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dw6r\" (UniqueName: \"kubernetes.io/projected/0eb4fa70-5bc2-4349-871f-1c095c1981ab-kube-api-access-8dw6r\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.793361 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-run-httpd\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.793452 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.793890 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-log-httpd\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.794012 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-run-httpd\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.803248 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.804186 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-config-data\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.806474 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-scripts\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.811604 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.812501 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.816499 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dw6r\" (UniqueName: \"kubernetes.io/projected/0eb4fa70-5bc2-4349-871f-1c095c1981ab-kube-api-access-8dw6r\") pod \"ceilometer-0\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " pod="openstack/ceilometer-0" Dec 15 12:40:05 crc kubenswrapper[4719]: I1215 12:40:05.863072 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:40:06 crc kubenswrapper[4719]: I1215 12:40:06.353243 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:06 crc kubenswrapper[4719]: I1215 12:40:06.975085 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-j88sc"] Dec 15 12:40:06 crc kubenswrapper[4719]: I1215 12:40:06.976396 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-j88sc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.004288 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-j88sc"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.116788 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbbzs\" (UniqueName: \"kubernetes.io/projected/42a85d8b-aca5-4aea-acfd-ee3ff752d272-kube-api-access-jbbzs\") pod \"nova-api-db-create-j88sc\" (UID: \"42a85d8b-aca5-4aea-acfd-ee3ff752d272\") " pod="openstack/nova-api-db-create-j88sc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.116933 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/42a85d8b-aca5-4aea-acfd-ee3ff752d272-operator-scripts\") pod \"nova-api-db-create-j88sc\" (UID: \"42a85d8b-aca5-4aea-acfd-ee3ff752d272\") " pod="openstack/nova-api-db-create-j88sc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.182315 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-bl7mv"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.183361 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bl7mv" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.200915 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bl7mv"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.210765 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-4957-account-create-update-r5lcc"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.212057 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0eb4fa70-5bc2-4349-871f-1c095c1981ab","Type":"ContainerStarted","Data":"40ac68518919ee1bc720f23e4aeec0385fa945abe444313e2ee3ae5814f5b751"} Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.212159 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4957-account-create-update-r5lcc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.213966 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.218548 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbbzs\" (UniqueName: \"kubernetes.io/projected/42a85d8b-aca5-4aea-acfd-ee3ff752d272-kube-api-access-jbbzs\") pod \"nova-api-db-create-j88sc\" (UID: \"42a85d8b-aca5-4aea-acfd-ee3ff752d272\") " pod="openstack/nova-api-db-create-j88sc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.218693 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/42a85d8b-aca5-4aea-acfd-ee3ff752d272-operator-scripts\") pod \"nova-api-db-create-j88sc\" (UID: \"42a85d8b-aca5-4aea-acfd-ee3ff752d272\") " pod="openstack/nova-api-db-create-j88sc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.244943 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-4957-account-create-update-r5lcc"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.256520 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/42a85d8b-aca5-4aea-acfd-ee3ff752d272-operator-scripts\") pod \"nova-api-db-create-j88sc\" (UID: \"42a85d8b-aca5-4aea-acfd-ee3ff752d272\") " pod="openstack/nova-api-db-create-j88sc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.272682 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbbzs\" (UniqueName: \"kubernetes.io/projected/42a85d8b-aca5-4aea-acfd-ee3ff752d272-kube-api-access-jbbzs\") pod \"nova-api-db-create-j88sc\" (UID: \"42a85d8b-aca5-4aea-acfd-ee3ff752d272\") " pod="openstack/nova-api-db-create-j88sc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.309388 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-j88sc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.322160 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5414c29-3e89-4a89-860d-9798ac533d81-operator-scripts\") pod \"nova-cell0-db-create-bl7mv\" (UID: \"b5414c29-3e89-4a89-860d-9798ac533d81\") " pod="openstack/nova-cell0-db-create-bl7mv" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.325374 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef50d20e-737c-486b-9759-08e17f34ce83-operator-scripts\") pod \"nova-api-4957-account-create-update-r5lcc\" (UID: \"ef50d20e-737c-486b-9759-08e17f34ce83\") " pod="openstack/nova-api-4957-account-create-update-r5lcc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.325527 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n85pt\" (UniqueName: \"kubernetes.io/projected/ef50d20e-737c-486b-9759-08e17f34ce83-kube-api-access-n85pt\") pod \"nova-api-4957-account-create-update-r5lcc\" (UID: \"ef50d20e-737c-486b-9759-08e17f34ce83\") " pod="openstack/nova-api-4957-account-create-update-r5lcc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.325583 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l62mq\" (UniqueName: \"kubernetes.io/projected/b5414c29-3e89-4a89-860d-9798ac533d81-kube-api-access-l62mq\") pod \"nova-cell0-db-create-bl7mv\" (UID: \"b5414c29-3e89-4a89-860d-9798ac533d81\") " pod="openstack/nova-cell0-db-create-bl7mv" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.386765 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-5w2hw"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.387869 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5w2hw" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.397001 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-30e9-account-create-update-rtsbx"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.398204 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.401054 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.406109 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5w2hw"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.414742 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-30e9-account-create-update-rtsbx"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.430375 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5414c29-3e89-4a89-860d-9798ac533d81-operator-scripts\") pod \"nova-cell0-db-create-bl7mv\" (UID: \"b5414c29-3e89-4a89-860d-9798ac533d81\") " pod="openstack/nova-cell0-db-create-bl7mv" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.430499 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef50d20e-737c-486b-9759-08e17f34ce83-operator-scripts\") pod \"nova-api-4957-account-create-update-r5lcc\" (UID: \"ef50d20e-737c-486b-9759-08e17f34ce83\") " pod="openstack/nova-api-4957-account-create-update-r5lcc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.430550 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n85pt\" (UniqueName: \"kubernetes.io/projected/ef50d20e-737c-486b-9759-08e17f34ce83-kube-api-access-n85pt\") pod \"nova-api-4957-account-create-update-r5lcc\" (UID: \"ef50d20e-737c-486b-9759-08e17f34ce83\") " pod="openstack/nova-api-4957-account-create-update-r5lcc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.430608 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l62mq\" (UniqueName: \"kubernetes.io/projected/b5414c29-3e89-4a89-860d-9798ac533d81-kube-api-access-l62mq\") pod \"nova-cell0-db-create-bl7mv\" (UID: \"b5414c29-3e89-4a89-860d-9798ac533d81\") " pod="openstack/nova-cell0-db-create-bl7mv" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.432147 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5414c29-3e89-4a89-860d-9798ac533d81-operator-scripts\") pod \"nova-cell0-db-create-bl7mv\" (UID: \"b5414c29-3e89-4a89-860d-9798ac533d81\") " pod="openstack/nova-cell0-db-create-bl7mv" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.448847 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef50d20e-737c-486b-9759-08e17f34ce83-operator-scripts\") pod \"nova-api-4957-account-create-update-r5lcc\" (UID: \"ef50d20e-737c-486b-9759-08e17f34ce83\") " pod="openstack/nova-api-4957-account-create-update-r5lcc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.493895 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n85pt\" (UniqueName: \"kubernetes.io/projected/ef50d20e-737c-486b-9759-08e17f34ce83-kube-api-access-n85pt\") pod \"nova-api-4957-account-create-update-r5lcc\" (UID: \"ef50d20e-737c-486b-9759-08e17f34ce83\") " pod="openstack/nova-api-4957-account-create-update-r5lcc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.494388 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l62mq\" (UniqueName: \"kubernetes.io/projected/b5414c29-3e89-4a89-860d-9798ac533d81-kube-api-access-l62mq\") pod \"nova-cell0-db-create-bl7mv\" (UID: \"b5414c29-3e89-4a89-860d-9798ac533d81\") " pod="openstack/nova-cell0-db-create-bl7mv" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.508700 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bl7mv" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.525533 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6dabdcb-160d-47b8-b616-bc13880955da" path="/var/lib/kubelet/pods/c6dabdcb-160d-47b8-b616-bc13880955da/volumes" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.533780 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73c6a473-c59b-481d-9047-a4006471710a-operator-scripts\") pod \"nova-cell0-30e9-account-create-update-rtsbx\" (UID: \"73c6a473-c59b-481d-9047-a4006471710a\") " pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.533868 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2825p\" (UniqueName: \"kubernetes.io/projected/e200319d-f1b8-421b-ba02-a9613dba7363-kube-api-access-2825p\") pod \"nova-cell1-db-create-5w2hw\" (UID: \"e200319d-f1b8-421b-ba02-a9613dba7363\") " pod="openstack/nova-cell1-db-create-5w2hw" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.533923 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e200319d-f1b8-421b-ba02-a9613dba7363-operator-scripts\") pod \"nova-cell1-db-create-5w2hw\" (UID: \"e200319d-f1b8-421b-ba02-a9613dba7363\") " pod="openstack/nova-cell1-db-create-5w2hw" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.533960 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjb6f\" (UniqueName: \"kubernetes.io/projected/73c6a473-c59b-481d-9047-a4006471710a-kube-api-access-xjb6f\") pod \"nova-cell0-30e9-account-create-update-rtsbx\" (UID: \"73c6a473-c59b-481d-9047-a4006471710a\") " pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.538355 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4957-account-create-update-r5lcc" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.595974 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc5fb48-jqf6b" podUID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.641393 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73c6a473-c59b-481d-9047-a4006471710a-operator-scripts\") pod \"nova-cell0-30e9-account-create-update-rtsbx\" (UID: \"73c6a473-c59b-481d-9047-a4006471710a\") " pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.640025 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73c6a473-c59b-481d-9047-a4006471710a-operator-scripts\") pod \"nova-cell0-30e9-account-create-update-rtsbx\" (UID: \"73c6a473-c59b-481d-9047-a4006471710a\") " pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.641521 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2825p\" (UniqueName: \"kubernetes.io/projected/e200319d-f1b8-421b-ba02-a9613dba7363-kube-api-access-2825p\") pod \"nova-cell1-db-create-5w2hw\" (UID: \"e200319d-f1b8-421b-ba02-a9613dba7363\") " pod="openstack/nova-cell1-db-create-5w2hw" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.655204 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e200319d-f1b8-421b-ba02-a9613dba7363-operator-scripts\") pod \"nova-cell1-db-create-5w2hw\" (UID: \"e200319d-f1b8-421b-ba02-a9613dba7363\") " pod="openstack/nova-cell1-db-create-5w2hw" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.655259 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjb6f\" (UniqueName: \"kubernetes.io/projected/73c6a473-c59b-481d-9047-a4006471710a-kube-api-access-xjb6f\") pod \"nova-cell0-30e9-account-create-update-rtsbx\" (UID: \"73c6a473-c59b-481d-9047-a4006471710a\") " pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.656400 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e200319d-f1b8-421b-ba02-a9613dba7363-operator-scripts\") pod \"nova-cell1-db-create-5w2hw\" (UID: \"e200319d-f1b8-421b-ba02-a9613dba7363\") " pod="openstack/nova-cell1-db-create-5w2hw" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.684615 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjb6f\" (UniqueName: \"kubernetes.io/projected/73c6a473-c59b-481d-9047-a4006471710a-kube-api-access-xjb6f\") pod \"nova-cell0-30e9-account-create-update-rtsbx\" (UID: \"73c6a473-c59b-481d-9047-a4006471710a\") " pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.705847 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-9663-account-create-update-fcjsm"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.707236 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9663-account-create-update-fcjsm" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.712491 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.717151 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2825p\" (UniqueName: \"kubernetes.io/projected/e200319d-f1b8-421b-ba02-a9613dba7363-kube-api-access-2825p\") pod \"nova-cell1-db-create-5w2hw\" (UID: \"e200319d-f1b8-421b-ba02-a9613dba7363\") " pod="openstack/nova-cell1-db-create-5w2hw" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.749130 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9663-account-create-update-fcjsm"] Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.769929 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5w2hw" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.793934 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.865766 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30b90f8c-d014-4952-b250-1aaaa52ea020-operator-scripts\") pod \"nova-cell1-9663-account-create-update-fcjsm\" (UID: \"30b90f8c-d014-4952-b250-1aaaa52ea020\") " pod="openstack/nova-cell1-9663-account-create-update-fcjsm" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.866173 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8r8c\" (UniqueName: \"kubernetes.io/projected/30b90f8c-d014-4952-b250-1aaaa52ea020-kube-api-access-h8r8c\") pod \"nova-cell1-9663-account-create-update-fcjsm\" (UID: \"30b90f8c-d014-4952-b250-1aaaa52ea020\") " pod="openstack/nova-cell1-9663-account-create-update-fcjsm" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.967389 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8r8c\" (UniqueName: \"kubernetes.io/projected/30b90f8c-d014-4952-b250-1aaaa52ea020-kube-api-access-h8r8c\") pod \"nova-cell1-9663-account-create-update-fcjsm\" (UID: \"30b90f8c-d014-4952-b250-1aaaa52ea020\") " pod="openstack/nova-cell1-9663-account-create-update-fcjsm" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.967531 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30b90f8c-d014-4952-b250-1aaaa52ea020-operator-scripts\") pod \"nova-cell1-9663-account-create-update-fcjsm\" (UID: \"30b90f8c-d014-4952-b250-1aaaa52ea020\") " pod="openstack/nova-cell1-9663-account-create-update-fcjsm" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.968276 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30b90f8c-d014-4952-b250-1aaaa52ea020-operator-scripts\") pod \"nova-cell1-9663-account-create-update-fcjsm\" (UID: \"30b90f8c-d014-4952-b250-1aaaa52ea020\") " pod="openstack/nova-cell1-9663-account-create-update-fcjsm" Dec 15 12:40:07 crc kubenswrapper[4719]: I1215 12:40:07.998669 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8r8c\" (UniqueName: \"kubernetes.io/projected/30b90f8c-d014-4952-b250-1aaaa52ea020-kube-api-access-h8r8c\") pod \"nova-cell1-9663-account-create-update-fcjsm\" (UID: \"30b90f8c-d014-4952-b250-1aaaa52ea020\") " pod="openstack/nova-cell1-9663-account-create-update-fcjsm" Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.020440 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-j88sc"] Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.052588 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9663-account-create-update-fcjsm" Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.244949 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-bl7mv"] Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.276028 4719 generic.go:334] "Generic (PLEG): container finished" podID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerID="e526058563f282b4d65e64a01944ac0f829154bf6a70328411350711beca2af4" exitCode=137 Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.276084 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6bcf58b6-vx99l" event={"ID":"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e","Type":"ContainerDied","Data":"e526058563f282b4d65e64a01944ac0f829154bf6a70328411350711beca2af4"} Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.276110 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6bcf58b6-vx99l" event={"ID":"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e","Type":"ContainerStarted","Data":"e1016fcc51f3a06758720611065b641214cb05d2109d1fc991685adda2a4994a"} Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.276125 4719 scope.go:117] "RemoveContainer" containerID="a515a01ab9e93d2a61d2349f044b8132b8de12998c6a016d169b104d98e71a08" Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.308581 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0eb4fa70-5bc2-4349-871f-1c095c1981ab","Type":"ContainerStarted","Data":"79401afe64fdd68b8c29e82df7fcd616c9058bf16c7991d6ce6e5a02f25e2cb8"} Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.328866 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-j88sc" event={"ID":"42a85d8b-aca5-4aea-acfd-ee3ff752d272","Type":"ContainerStarted","Data":"3a5c9b337625fab96fa2dfc925246afcd2685c25fb3485109cc158e3fa331781"} Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.435346 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-4957-account-create-update-r5lcc"] Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.569368 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5w2hw"] Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.669188 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-30e9-account-create-update-rtsbx"] Dec 15 12:40:08 crc kubenswrapper[4719]: I1215 12:40:08.925658 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9663-account-create-update-fcjsm"] Dec 15 12:40:09 crc kubenswrapper[4719]: W1215 12:40:09.001780 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30b90f8c_d014_4952_b250_1aaaa52ea020.slice/crio-56200b7cbbdc17358d9bd79b837970bf88849b5afc68da7112d034dbcbbc4239 WatchSource:0}: Error finding container 56200b7cbbdc17358d9bd79b837970bf88849b5afc68da7112d034dbcbbc4239: Status 404 returned error can't find the container with id 56200b7cbbdc17358d9bd79b837970bf88849b5afc68da7112d034dbcbbc4239 Dec 15 12:40:09 crc kubenswrapper[4719]: I1215 12:40:09.384762 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0eb4fa70-5bc2-4349-871f-1c095c1981ab","Type":"ContainerStarted","Data":"816196fa0dc9b093ce389dd7c9fd108bd935ba0a54cfe31b448d617e7b882cb4"} Dec 15 12:40:09 crc kubenswrapper[4719]: I1215 12:40:09.398799 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-4957-account-create-update-r5lcc" event={"ID":"ef50d20e-737c-486b-9759-08e17f34ce83","Type":"ContainerStarted","Data":"981a44ffc90bddfd24e0d3b4ccd1d971c4ee4145e261511848d4031136990fa9"} Dec 15 12:40:09 crc kubenswrapper[4719]: I1215 12:40:09.418585 4719 generic.go:334] "Generic (PLEG): container finished" podID="42a85d8b-aca5-4aea-acfd-ee3ff752d272" containerID="71ab1ddbafb71d195095c6f6fe0f9e84bfb738bddf836284b678eb37ecb0effb" exitCode=0 Dec 15 12:40:09 crc kubenswrapper[4719]: I1215 12:40:09.418916 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-j88sc" event={"ID":"42a85d8b-aca5-4aea-acfd-ee3ff752d272","Type":"ContainerDied","Data":"71ab1ddbafb71d195095c6f6fe0f9e84bfb738bddf836284b678eb37ecb0effb"} Dec 15 12:40:09 crc kubenswrapper[4719]: I1215 12:40:09.429039 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5w2hw" event={"ID":"e200319d-f1b8-421b-ba02-a9613dba7363","Type":"ContainerStarted","Data":"b044a211eb57553d004f01dd6f91ea723e16ab010e1f45716bb4ca47cdc0537b"} Dec 15 12:40:09 crc kubenswrapper[4719]: I1215 12:40:09.452113 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bl7mv" event={"ID":"b5414c29-3e89-4a89-860d-9798ac533d81","Type":"ContainerStarted","Data":"f08747ef15487cdbde3a0a58138ca5bc9934fbcd5aace0cff6a4e4512a4f8e46"} Dec 15 12:40:09 crc kubenswrapper[4719]: I1215 12:40:09.452156 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bl7mv" event={"ID":"b5414c29-3e89-4a89-860d-9798ac533d81","Type":"ContainerStarted","Data":"c0429d054efa772ec7c4bebba8e233c74416d2e6859ba5ab81ef222cbeacaff4"} Dec 15 12:40:09 crc kubenswrapper[4719]: I1215 12:40:09.456425 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9663-account-create-update-fcjsm" event={"ID":"30b90f8c-d014-4952-b250-1aaaa52ea020","Type":"ContainerStarted","Data":"56200b7cbbdc17358d9bd79b837970bf88849b5afc68da7112d034dbcbbc4239"} Dec 15 12:40:09 crc kubenswrapper[4719]: I1215 12:40:09.469092 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" event={"ID":"73c6a473-c59b-481d-9047-a4006471710a","Type":"ContainerStarted","Data":"e4b500f857b493894807194df78a150b71359b10990f68f275be346607752207"} Dec 15 12:40:09 crc kubenswrapper[4719]: I1215 12:40:09.521937 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-bl7mv" podStartSLOduration=2.52192111 podStartE2EDuration="2.52192111s" podCreationTimestamp="2025-12-15 12:40:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:40:09.482251387 +0000 UTC m=+1370.424544417" watchObservedRunningTime="2025-12-15 12:40:09.52192111 +0000 UTC m=+1370.464214140" Dec 15 12:40:10 crc kubenswrapper[4719]: I1215 12:40:10.497753 4719 generic.go:334] "Generic (PLEG): container finished" podID="e200319d-f1b8-421b-ba02-a9613dba7363" containerID="3ce57c3590ea2f21308b981409e71b532cb255058381148ed40ad0853227af38" exitCode=0 Dec 15 12:40:10 crc kubenswrapper[4719]: I1215 12:40:10.498283 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5w2hw" event={"ID":"e200319d-f1b8-421b-ba02-a9613dba7363","Type":"ContainerDied","Data":"3ce57c3590ea2f21308b981409e71b532cb255058381148ed40ad0853227af38"} Dec 15 12:40:10 crc kubenswrapper[4719]: I1215 12:40:10.523601 4719 generic.go:334] "Generic (PLEG): container finished" podID="b5414c29-3e89-4a89-860d-9798ac533d81" containerID="f08747ef15487cdbde3a0a58138ca5bc9934fbcd5aace0cff6a4e4512a4f8e46" exitCode=0 Dec 15 12:40:10 crc kubenswrapper[4719]: I1215 12:40:10.523707 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bl7mv" event={"ID":"b5414c29-3e89-4a89-860d-9798ac533d81","Type":"ContainerDied","Data":"f08747ef15487cdbde3a0a58138ca5bc9934fbcd5aace0cff6a4e4512a4f8e46"} Dec 15 12:40:10 crc kubenswrapper[4719]: I1215 12:40:10.547402 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9663-account-create-update-fcjsm" event={"ID":"30b90f8c-d014-4952-b250-1aaaa52ea020","Type":"ContainerStarted","Data":"9054fde60087e4364449be7dabef9313151a2c1a3c452049aba7f8f452554efd"} Dec 15 12:40:10 crc kubenswrapper[4719]: I1215 12:40:10.550054 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" event={"ID":"73c6a473-c59b-481d-9047-a4006471710a","Type":"ContainerStarted","Data":"47968e61c716203e7cb65307363c747435b970d7135f2b99b11c0f24f0411100"} Dec 15 12:40:10 crc kubenswrapper[4719]: I1215 12:40:10.557749 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0eb4fa70-5bc2-4349-871f-1c095c1981ab","Type":"ContainerStarted","Data":"308fa4c38c880ecdaa47deefd03e442682aaf080c5eae1b938513752f5fba420"} Dec 15 12:40:10 crc kubenswrapper[4719]: I1215 12:40:10.560402 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-4957-account-create-update-r5lcc" event={"ID":"ef50d20e-737c-486b-9759-08e17f34ce83","Type":"ContainerStarted","Data":"9da848555b8101c4361795c4e608af88e63cc10e75dbbfc98af7be184d1ed1f9"} Dec 15 12:40:10 crc kubenswrapper[4719]: I1215 12:40:10.632053 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-4957-account-create-update-r5lcc" podStartSLOduration=3.632030929 podStartE2EDuration="3.632030929s" podCreationTimestamp="2025-12-15 12:40:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:40:10.594336187 +0000 UTC m=+1371.536629207" watchObservedRunningTime="2025-12-15 12:40:10.632030929 +0000 UTC m=+1371.574323959" Dec 15 12:40:10 crc kubenswrapper[4719]: I1215 12:40:10.665120 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" podStartSLOduration=3.665099655 podStartE2EDuration="3.665099655s" podCreationTimestamp="2025-12-15 12:40:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:40:10.60910338 +0000 UTC m=+1371.551396410" watchObservedRunningTime="2025-12-15 12:40:10.665099655 +0000 UTC m=+1371.607392685" Dec 15 12:40:10 crc kubenswrapper[4719]: E1215 12:40:10.725780 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30b90f8c_d014_4952_b250_1aaaa52ea020.slice/crio-conmon-9054fde60087e4364449be7dabef9313151a2c1a3c452049aba7f8f452554efd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode200319d_f1b8_421b_ba02_a9613dba7363.slice/crio-conmon-3ce57c3590ea2f21308b981409e71b532cb255058381148ed40ad0853227af38.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef50d20e_737c_486b_9759_08e17f34ce83.slice/crio-9da848555b8101c4361795c4e608af88e63cc10e75dbbfc98af7be184d1ed1f9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73c6a473_c59b_481d_9047_a4006471710a.slice/crio-conmon-47968e61c716203e7cb65307363c747435b970d7135f2b99b11c0f24f0411100.scope\": RecentStats: unable to find data in memory cache]" Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.038591 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-j88sc" Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.164739 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/42a85d8b-aca5-4aea-acfd-ee3ff752d272-operator-scripts\") pod \"42a85d8b-aca5-4aea-acfd-ee3ff752d272\" (UID: \"42a85d8b-aca5-4aea-acfd-ee3ff752d272\") " Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.165197 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42a85d8b-aca5-4aea-acfd-ee3ff752d272-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "42a85d8b-aca5-4aea-acfd-ee3ff752d272" (UID: "42a85d8b-aca5-4aea-acfd-ee3ff752d272"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.165349 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbbzs\" (UniqueName: \"kubernetes.io/projected/42a85d8b-aca5-4aea-acfd-ee3ff752d272-kube-api-access-jbbzs\") pod \"42a85d8b-aca5-4aea-acfd-ee3ff752d272\" (UID: \"42a85d8b-aca5-4aea-acfd-ee3ff752d272\") " Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.165938 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/42a85d8b-aca5-4aea-acfd-ee3ff752d272-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.191155 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a85d8b-aca5-4aea-acfd-ee3ff752d272-kube-api-access-jbbzs" (OuterVolumeSpecName: "kube-api-access-jbbzs") pod "42a85d8b-aca5-4aea-acfd-ee3ff752d272" (UID: "42a85d8b-aca5-4aea-acfd-ee3ff752d272"). InnerVolumeSpecName "kube-api-access-jbbzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.268810 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbbzs\" (UniqueName: \"kubernetes.io/projected/42a85d8b-aca5-4aea-acfd-ee3ff752d272-kube-api-access-jbbzs\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.633289 4719 generic.go:334] "Generic (PLEG): container finished" podID="30b90f8c-d014-4952-b250-1aaaa52ea020" containerID="9054fde60087e4364449be7dabef9313151a2c1a3c452049aba7f8f452554efd" exitCode=0 Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.633604 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9663-account-create-update-fcjsm" event={"ID":"30b90f8c-d014-4952-b250-1aaaa52ea020","Type":"ContainerDied","Data":"9054fde60087e4364449be7dabef9313151a2c1a3c452049aba7f8f452554efd"} Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.650290 4719 generic.go:334] "Generic (PLEG): container finished" podID="73c6a473-c59b-481d-9047-a4006471710a" containerID="47968e61c716203e7cb65307363c747435b970d7135f2b99b11c0f24f0411100" exitCode=0 Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.650351 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" event={"ID":"73c6a473-c59b-481d-9047-a4006471710a","Type":"ContainerDied","Data":"47968e61c716203e7cb65307363c747435b970d7135f2b99b11c0f24f0411100"} Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.687533 4719 generic.go:334] "Generic (PLEG): container finished" podID="ef50d20e-737c-486b-9759-08e17f34ce83" containerID="9da848555b8101c4361795c4e608af88e63cc10e75dbbfc98af7be184d1ed1f9" exitCode=0 Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.687632 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-4957-account-create-update-r5lcc" event={"ID":"ef50d20e-737c-486b-9759-08e17f34ce83","Type":"ContainerDied","Data":"9da848555b8101c4361795c4e608af88e63cc10e75dbbfc98af7be184d1ed1f9"} Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.714043 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-j88sc" Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.722193 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-j88sc" event={"ID":"42a85d8b-aca5-4aea-acfd-ee3ff752d272","Type":"ContainerDied","Data":"3a5c9b337625fab96fa2dfc925246afcd2685c25fb3485109cc158e3fa331781"} Dec 15 12:40:11 crc kubenswrapper[4719]: I1215 12:40:11.722341 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a5c9b337625fab96fa2dfc925246afcd2685c25fb3485109cc158e3fa331781" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.456341 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9663-account-create-update-fcjsm" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.460388 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5w2hw" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.474542 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bl7mv" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.627333 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e200319d-f1b8-421b-ba02-a9613dba7363-operator-scripts\") pod \"e200319d-f1b8-421b-ba02-a9613dba7363\" (UID: \"e200319d-f1b8-421b-ba02-a9613dba7363\") " Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.627754 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l62mq\" (UniqueName: \"kubernetes.io/projected/b5414c29-3e89-4a89-860d-9798ac533d81-kube-api-access-l62mq\") pod \"b5414c29-3e89-4a89-860d-9798ac533d81\" (UID: \"b5414c29-3e89-4a89-860d-9798ac533d81\") " Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.627820 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8r8c\" (UniqueName: \"kubernetes.io/projected/30b90f8c-d014-4952-b250-1aaaa52ea020-kube-api-access-h8r8c\") pod \"30b90f8c-d014-4952-b250-1aaaa52ea020\" (UID: \"30b90f8c-d014-4952-b250-1aaaa52ea020\") " Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.627844 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30b90f8c-d014-4952-b250-1aaaa52ea020-operator-scripts\") pod \"30b90f8c-d014-4952-b250-1aaaa52ea020\" (UID: \"30b90f8c-d014-4952-b250-1aaaa52ea020\") " Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.627926 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2825p\" (UniqueName: \"kubernetes.io/projected/e200319d-f1b8-421b-ba02-a9613dba7363-kube-api-access-2825p\") pod \"e200319d-f1b8-421b-ba02-a9613dba7363\" (UID: \"e200319d-f1b8-421b-ba02-a9613dba7363\") " Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.627975 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5414c29-3e89-4a89-860d-9798ac533d81-operator-scripts\") pod \"b5414c29-3e89-4a89-860d-9798ac533d81\" (UID: \"b5414c29-3e89-4a89-860d-9798ac533d81\") " Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.628148 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e200319d-f1b8-421b-ba02-a9613dba7363-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e200319d-f1b8-421b-ba02-a9613dba7363" (UID: "e200319d-f1b8-421b-ba02-a9613dba7363"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.628383 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e200319d-f1b8-421b-ba02-a9613dba7363-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.628430 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5414c29-3e89-4a89-860d-9798ac533d81-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5414c29-3e89-4a89-860d-9798ac533d81" (UID: "b5414c29-3e89-4a89-860d-9798ac533d81"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.628455 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b90f8c-d014-4952-b250-1aaaa52ea020-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "30b90f8c-d014-4952-b250-1aaaa52ea020" (UID: "30b90f8c-d014-4952-b250-1aaaa52ea020"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.636036 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5414c29-3e89-4a89-860d-9798ac533d81-kube-api-access-l62mq" (OuterVolumeSpecName: "kube-api-access-l62mq") pod "b5414c29-3e89-4a89-860d-9798ac533d81" (UID: "b5414c29-3e89-4a89-860d-9798ac533d81"). InnerVolumeSpecName "kube-api-access-l62mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.637608 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e200319d-f1b8-421b-ba02-a9613dba7363-kube-api-access-2825p" (OuterVolumeSpecName: "kube-api-access-2825p") pod "e200319d-f1b8-421b-ba02-a9613dba7363" (UID: "e200319d-f1b8-421b-ba02-a9613dba7363"). InnerVolumeSpecName "kube-api-access-2825p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.644161 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30b90f8c-d014-4952-b250-1aaaa52ea020-kube-api-access-h8r8c" (OuterVolumeSpecName: "kube-api-access-h8r8c") pod "30b90f8c-d014-4952-b250-1aaaa52ea020" (UID: "30b90f8c-d014-4952-b250-1aaaa52ea020"). InnerVolumeSpecName "kube-api-access-h8r8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.725403 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9663-account-create-update-fcjsm" event={"ID":"30b90f8c-d014-4952-b250-1aaaa52ea020","Type":"ContainerDied","Data":"56200b7cbbdc17358d9bd79b837970bf88849b5afc68da7112d034dbcbbc4239"} Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.725765 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56200b7cbbdc17358d9bd79b837970bf88849b5afc68da7112d034dbcbbc4239" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.725740 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9663-account-create-update-fcjsm" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.729601 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l62mq\" (UniqueName: \"kubernetes.io/projected/b5414c29-3e89-4a89-860d-9798ac533d81-kube-api-access-l62mq\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.730566 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8r8c\" (UniqueName: \"kubernetes.io/projected/30b90f8c-d014-4952-b250-1aaaa52ea020-kube-api-access-h8r8c\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.730702 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30b90f8c-d014-4952-b250-1aaaa52ea020-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.730773 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2825p\" (UniqueName: \"kubernetes.io/projected/e200319d-f1b8-421b-ba02-a9613dba7363-kube-api-access-2825p\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.730828 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5414c29-3e89-4a89-860d-9798ac533d81-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.743033 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0eb4fa70-5bc2-4349-871f-1c095c1981ab","Type":"ContainerStarted","Data":"bd78bb3a6074a09b2bc0b5554ff22dc2a6e0e0cc0e2e9ff6613d3118120e0e78"} Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.743952 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.751718 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5w2hw" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.752263 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5w2hw" event={"ID":"e200319d-f1b8-421b-ba02-a9613dba7363","Type":"ContainerDied","Data":"b044a211eb57553d004f01dd6f91ea723e16ab010e1f45716bb4ca47cdc0537b"} Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.752314 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b044a211eb57553d004f01dd6f91ea723e16ab010e1f45716bb4ca47cdc0537b" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.760575 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-bl7mv" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.760670 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-bl7mv" event={"ID":"b5414c29-3e89-4a89-860d-9798ac533d81","Type":"ContainerDied","Data":"c0429d054efa772ec7c4bebba8e233c74416d2e6859ba5ab81ef222cbeacaff4"} Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.761035 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0429d054efa772ec7c4bebba8e233c74416d2e6859ba5ab81ef222cbeacaff4" Dec 15 12:40:12 crc kubenswrapper[4719]: I1215 12:40:12.772095 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.571343367 podStartE2EDuration="7.772070231s" podCreationTimestamp="2025-12-15 12:40:05 +0000 UTC" firstStartedPulling="2025-12-15 12:40:06.362218035 +0000 UTC m=+1367.304511075" lastFinishedPulling="2025-12-15 12:40:11.562944909 +0000 UTC m=+1372.505237939" observedRunningTime="2025-12-15 12:40:12.766264969 +0000 UTC m=+1373.708557999" watchObservedRunningTime="2025-12-15 12:40:12.772070231 +0000 UTC m=+1373.714363261" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.072931 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.141165 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjb6f\" (UniqueName: \"kubernetes.io/projected/73c6a473-c59b-481d-9047-a4006471710a-kube-api-access-xjb6f\") pod \"73c6a473-c59b-481d-9047-a4006471710a\" (UID: \"73c6a473-c59b-481d-9047-a4006471710a\") " Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.141273 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73c6a473-c59b-481d-9047-a4006471710a-operator-scripts\") pod \"73c6a473-c59b-481d-9047-a4006471710a\" (UID: \"73c6a473-c59b-481d-9047-a4006471710a\") " Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.142501 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73c6a473-c59b-481d-9047-a4006471710a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "73c6a473-c59b-481d-9047-a4006471710a" (UID: "73c6a473-c59b-481d-9047-a4006471710a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.145508 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4957-account-create-update-r5lcc" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.148982 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73c6a473-c59b-481d-9047-a4006471710a-kube-api-access-xjb6f" (OuterVolumeSpecName: "kube-api-access-xjb6f") pod "73c6a473-c59b-481d-9047-a4006471710a" (UID: "73c6a473-c59b-481d-9047-a4006471710a"). InnerVolumeSpecName "kube-api-access-xjb6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.243562 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n85pt\" (UniqueName: \"kubernetes.io/projected/ef50d20e-737c-486b-9759-08e17f34ce83-kube-api-access-n85pt\") pod \"ef50d20e-737c-486b-9759-08e17f34ce83\" (UID: \"ef50d20e-737c-486b-9759-08e17f34ce83\") " Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.243690 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef50d20e-737c-486b-9759-08e17f34ce83-operator-scripts\") pod \"ef50d20e-737c-486b-9759-08e17f34ce83\" (UID: \"ef50d20e-737c-486b-9759-08e17f34ce83\") " Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.244082 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjb6f\" (UniqueName: \"kubernetes.io/projected/73c6a473-c59b-481d-9047-a4006471710a-kube-api-access-xjb6f\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.244095 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73c6a473-c59b-481d-9047-a4006471710a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.244399 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef50d20e-737c-486b-9759-08e17f34ce83-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ef50d20e-737c-486b-9759-08e17f34ce83" (UID: "ef50d20e-737c-486b-9759-08e17f34ce83"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.249981 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef50d20e-737c-486b-9759-08e17f34ce83-kube-api-access-n85pt" (OuterVolumeSpecName: "kube-api-access-n85pt") pod "ef50d20e-737c-486b-9759-08e17f34ce83" (UID: "ef50d20e-737c-486b-9759-08e17f34ce83"). InnerVolumeSpecName "kube-api-access-n85pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.346228 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n85pt\" (UniqueName: \"kubernetes.io/projected/ef50d20e-737c-486b-9759-08e17f34ce83-kube-api-access-n85pt\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.346262 4719 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef50d20e-737c-486b-9759-08e17f34ce83-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.808411 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" event={"ID":"73c6a473-c59b-481d-9047-a4006471710a","Type":"ContainerDied","Data":"e4b500f857b493894807194df78a150b71359b10990f68f275be346607752207"} Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.808670 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4b500f857b493894807194df78a150b71359b10990f68f275be346607752207" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.808723 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30e9-account-create-update-rtsbx" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.811720 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-4957-account-create-update-r5lcc" Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.811776 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-4957-account-create-update-r5lcc" event={"ID":"ef50d20e-737c-486b-9759-08e17f34ce83","Type":"ContainerDied","Data":"981a44ffc90bddfd24e0d3b4ccd1d971c4ee4145e261511848d4031136990fa9"} Dec 15 12:40:13 crc kubenswrapper[4719]: I1215 12:40:13.811801 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="981a44ffc90bddfd24e0d3b4ccd1d971c4ee4145e261511848d4031136990fa9" Dec 15 12:40:14 crc kubenswrapper[4719]: I1215 12:40:14.477638 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:15 crc kubenswrapper[4719]: I1215 12:40:15.827708 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="ceilometer-central-agent" containerID="cri-o://79401afe64fdd68b8c29e82df7fcd616c9058bf16c7991d6ce6e5a02f25e2cb8" gracePeriod=30 Dec 15 12:40:15 crc kubenswrapper[4719]: I1215 12:40:15.827755 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="sg-core" containerID="cri-o://308fa4c38c880ecdaa47deefd03e442682aaf080c5eae1b938513752f5fba420" gracePeriod=30 Dec 15 12:40:15 crc kubenswrapper[4719]: I1215 12:40:15.827765 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="proxy-httpd" containerID="cri-o://bd78bb3a6074a09b2bc0b5554ff22dc2a6e0e0cc0e2e9ff6613d3118120e0e78" gracePeriod=30 Dec 15 12:40:15 crc kubenswrapper[4719]: I1215 12:40:15.827768 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="ceilometer-notification-agent" containerID="cri-o://816196fa0dc9b093ce389dd7c9fd108bd935ba0a54cfe31b448d617e7b882cb4" gracePeriod=30 Dec 15 12:40:16 crc kubenswrapper[4719]: I1215 12:40:16.838194 4719 generic.go:334] "Generic (PLEG): container finished" podID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerID="bd78bb3a6074a09b2bc0b5554ff22dc2a6e0e0cc0e2e9ff6613d3118120e0e78" exitCode=0 Dec 15 12:40:16 crc kubenswrapper[4719]: I1215 12:40:16.838537 4719 generic.go:334] "Generic (PLEG): container finished" podID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerID="308fa4c38c880ecdaa47deefd03e442682aaf080c5eae1b938513752f5fba420" exitCode=2 Dec 15 12:40:16 crc kubenswrapper[4719]: I1215 12:40:16.838549 4719 generic.go:334] "Generic (PLEG): container finished" podID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerID="816196fa0dc9b093ce389dd7c9fd108bd935ba0a54cfe31b448d617e7b882cb4" exitCode=0 Dec 15 12:40:16 crc kubenswrapper[4719]: I1215 12:40:16.838572 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0eb4fa70-5bc2-4349-871f-1c095c1981ab","Type":"ContainerDied","Data":"bd78bb3a6074a09b2bc0b5554ff22dc2a6e0e0cc0e2e9ff6613d3118120e0e78"} Dec 15 12:40:16 crc kubenswrapper[4719]: I1215 12:40:16.838601 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0eb4fa70-5bc2-4349-871f-1c095c1981ab","Type":"ContainerDied","Data":"308fa4c38c880ecdaa47deefd03e442682aaf080c5eae1b938513752f5fba420"} Dec 15 12:40:16 crc kubenswrapper[4719]: I1215 12:40:16.838617 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0eb4fa70-5bc2-4349-871f-1c095c1981ab","Type":"ContainerDied","Data":"816196fa0dc9b093ce389dd7c9fd108bd935ba0a54cfe31b448d617e7b882cb4"} Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.389480 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.389533 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.391072 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.575097 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6bc5fb48-jqf6b" podUID="f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.666421 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x6vc4"] Dec 15 12:40:17 crc kubenswrapper[4719]: E1215 12:40:17.666864 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b90f8c-d014-4952-b250-1aaaa52ea020" containerName="mariadb-account-create-update" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.666881 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b90f8c-d014-4952-b250-1aaaa52ea020" containerName="mariadb-account-create-update" Dec 15 12:40:17 crc kubenswrapper[4719]: E1215 12:40:17.666893 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a85d8b-aca5-4aea-acfd-ee3ff752d272" containerName="mariadb-database-create" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.666900 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a85d8b-aca5-4aea-acfd-ee3ff752d272" containerName="mariadb-database-create" Dec 15 12:40:17 crc kubenswrapper[4719]: E1215 12:40:17.666916 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73c6a473-c59b-481d-9047-a4006471710a" containerName="mariadb-account-create-update" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.666924 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="73c6a473-c59b-481d-9047-a4006471710a" containerName="mariadb-account-create-update" Dec 15 12:40:17 crc kubenswrapper[4719]: E1215 12:40:17.666943 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5414c29-3e89-4a89-860d-9798ac533d81" containerName="mariadb-database-create" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.666949 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5414c29-3e89-4a89-860d-9798ac533d81" containerName="mariadb-database-create" Dec 15 12:40:17 crc kubenswrapper[4719]: E1215 12:40:17.666956 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef50d20e-737c-486b-9759-08e17f34ce83" containerName="mariadb-account-create-update" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.666962 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef50d20e-737c-486b-9759-08e17f34ce83" containerName="mariadb-account-create-update" Dec 15 12:40:17 crc kubenswrapper[4719]: E1215 12:40:17.666976 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e200319d-f1b8-421b-ba02-a9613dba7363" containerName="mariadb-database-create" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.666984 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="e200319d-f1b8-421b-ba02-a9613dba7363" containerName="mariadb-database-create" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.667160 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="30b90f8c-d014-4952-b250-1aaaa52ea020" containerName="mariadb-account-create-update" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.667176 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="42a85d8b-aca5-4aea-acfd-ee3ff752d272" containerName="mariadb-database-create" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.667188 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="e200319d-f1b8-421b-ba02-a9613dba7363" containerName="mariadb-database-create" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.667200 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="73c6a473-c59b-481d-9047-a4006471710a" containerName="mariadb-account-create-update" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.667211 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5414c29-3e89-4a89-860d-9798ac533d81" containerName="mariadb-database-create" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.667223 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef50d20e-737c-486b-9759-08e17f34ce83" containerName="mariadb-account-create-update" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.667751 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.670051 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gj44p" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.670090 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.670435 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.716605 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x6vc4"] Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.731838 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-scripts\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.731895 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.731977 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-config-data\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.732057 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vcqr\" (UniqueName: \"kubernetes.io/projected/f1f092f5-6e7a-408d-8985-0a900417af4e-kube-api-access-7vcqr\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.833842 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vcqr\" (UniqueName: \"kubernetes.io/projected/f1f092f5-6e7a-408d-8985-0a900417af4e-kube-api-access-7vcqr\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.833917 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-scripts\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.833954 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.834031 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-config-data\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.841308 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-scripts\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.843482 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.848685 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-config-data\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.853986 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vcqr\" (UniqueName: \"kubernetes.io/projected/f1f092f5-6e7a-408d-8985-0a900417af4e-kube-api-access-7vcqr\") pod \"nova-cell0-conductor-db-sync-x6vc4\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:17 crc kubenswrapper[4719]: I1215 12:40:17.986651 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:18 crc kubenswrapper[4719]: I1215 12:40:18.659807 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x6vc4"] Dec 15 12:40:18 crc kubenswrapper[4719]: I1215 12:40:18.867093 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x6vc4" event={"ID":"f1f092f5-6e7a-408d-8985-0a900417af4e","Type":"ContainerStarted","Data":"d70a8f04742d716a3313cc22d23a740a92bba16013420a7fc2d1283890491f53"} Dec 15 12:40:19 crc kubenswrapper[4719]: I1215 12:40:19.884594 4719 generic.go:334] "Generic (PLEG): container finished" podID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerID="79401afe64fdd68b8c29e82df7fcd616c9058bf16c7991d6ce6e5a02f25e2cb8" exitCode=0 Dec 15 12:40:19 crc kubenswrapper[4719]: I1215 12:40:19.885183 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0eb4fa70-5bc2-4349-871f-1c095c1981ab","Type":"ContainerDied","Data":"79401afe64fdd68b8c29e82df7fcd616c9058bf16c7991d6ce6e5a02f25e2cb8"} Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.168225 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.284447 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-scripts\") pod \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.284543 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-sg-core-conf-yaml\") pod \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.284588 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-combined-ca-bundle\") pod \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.284624 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dw6r\" (UniqueName: \"kubernetes.io/projected/0eb4fa70-5bc2-4349-871f-1c095c1981ab-kube-api-access-8dw6r\") pod \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.284669 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-run-httpd\") pod \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.284721 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-config-data\") pod \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.284749 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-ceilometer-tls-certs\") pod \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.284817 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-log-httpd\") pod \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\" (UID: \"0eb4fa70-5bc2-4349-871f-1c095c1981ab\") " Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.285375 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0eb4fa70-5bc2-4349-871f-1c095c1981ab" (UID: "0eb4fa70-5bc2-4349-871f-1c095c1981ab"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.285554 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0eb4fa70-5bc2-4349-871f-1c095c1981ab" (UID: "0eb4fa70-5bc2-4349-871f-1c095c1981ab"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.296033 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eb4fa70-5bc2-4349-871f-1c095c1981ab-kube-api-access-8dw6r" (OuterVolumeSpecName: "kube-api-access-8dw6r") pod "0eb4fa70-5bc2-4349-871f-1c095c1981ab" (UID: "0eb4fa70-5bc2-4349-871f-1c095c1981ab"). InnerVolumeSpecName "kube-api-access-8dw6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.296165 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-scripts" (OuterVolumeSpecName: "scripts") pod "0eb4fa70-5bc2-4349-871f-1c095c1981ab" (UID: "0eb4fa70-5bc2-4349-871f-1c095c1981ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.317057 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0eb4fa70-5bc2-4349-871f-1c095c1981ab" (UID: "0eb4fa70-5bc2-4349-871f-1c095c1981ab"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.368334 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0eb4fa70-5bc2-4349-871f-1c095c1981ab" (UID: "0eb4fa70-5bc2-4349-871f-1c095c1981ab"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.375657 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0eb4fa70-5bc2-4349-871f-1c095c1981ab" (UID: "0eb4fa70-5bc2-4349-871f-1c095c1981ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.387318 4719 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.387343 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.387355 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dw6r\" (UniqueName: \"kubernetes.io/projected/0eb4fa70-5bc2-4349-871f-1c095c1981ab-kube-api-access-8dw6r\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.387368 4719 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.387379 4719 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.387392 4719 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0eb4fa70-5bc2-4349-871f-1c095c1981ab-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.387403 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.413750 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-config-data" (OuterVolumeSpecName: "config-data") pod "0eb4fa70-5bc2-4349-871f-1c095c1981ab" (UID: "0eb4fa70-5bc2-4349-871f-1c095c1981ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.488931 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0eb4fa70-5bc2-4349-871f-1c095c1981ab-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.896896 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0eb4fa70-5bc2-4349-871f-1c095c1981ab","Type":"ContainerDied","Data":"40ac68518919ee1bc720f23e4aeec0385fa945abe444313e2ee3ae5814f5b751"} Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.896953 4719 scope.go:117] "RemoveContainer" containerID="bd78bb3a6074a09b2bc0b5554ff22dc2a6e0e0cc0e2e9ff6613d3118120e0e78" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.896959 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.931591 4719 scope.go:117] "RemoveContainer" containerID="308fa4c38c880ecdaa47deefd03e442682aaf080c5eae1b938513752f5fba420" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.938172 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.956621 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.986337 4719 scope.go:117] "RemoveContainer" containerID="816196fa0dc9b093ce389dd7c9fd108bd935ba0a54cfe31b448d617e7b882cb4" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.997064 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:20 crc kubenswrapper[4719]: E1215 12:40:20.997580 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="sg-core" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.997599 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="sg-core" Dec 15 12:40:20 crc kubenswrapper[4719]: E1215 12:40:20.997615 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="ceilometer-notification-agent" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.997622 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="ceilometer-notification-agent" Dec 15 12:40:20 crc kubenswrapper[4719]: E1215 12:40:20.997633 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="ceilometer-central-agent" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.997639 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="ceilometer-central-agent" Dec 15 12:40:20 crc kubenswrapper[4719]: E1215 12:40:20.997651 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="proxy-httpd" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.997659 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="proxy-httpd" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.997843 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="proxy-httpd" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.997923 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="sg-core" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.997932 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="ceilometer-notification-agent" Dec 15 12:40:20 crc kubenswrapper[4719]: I1215 12:40:20.997944 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" containerName="ceilometer-central-agent" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.000167 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.007400 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.007618 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.007767 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.009122 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.022781 4719 scope.go:117] "RemoveContainer" containerID="79401afe64fdd68b8c29e82df7fcd616c9058bf16c7991d6ce6e5a02f25e2cb8" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.103976 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crlrx\" (UniqueName: \"kubernetes.io/projected/de640a25-d092-4c73-bb2f-c7a21993261e-kube-api-access-crlrx\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.104048 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.104077 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-run-httpd\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.104255 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-config-data\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.104337 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-scripts\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.104480 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-log-httpd\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.104602 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.104674 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.206073 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-run-httpd\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.206168 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-config-data\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.206208 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-scripts\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.206266 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-log-httpd\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.206321 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.206358 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.206382 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crlrx\" (UniqueName: \"kubernetes.io/projected/de640a25-d092-4c73-bb2f-c7a21993261e-kube-api-access-crlrx\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.206436 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.207279 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-run-httpd\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.212240 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-scripts\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.212410 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-config-data\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.213215 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.226160 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.227054 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-log-httpd\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.231003 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.233919 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crlrx\" (UniqueName: \"kubernetes.io/projected/de640a25-d092-4c73-bb2f-c7a21993261e-kube-api-access-crlrx\") pod \"ceilometer-0\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.323481 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.501108 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0eb4fa70-5bc2-4349-871f-1c095c1981ab" path="/var/lib/kubelet/pods/0eb4fa70-5bc2-4349-871f-1c095c1981ab/volumes" Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.811110 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:40:21 crc kubenswrapper[4719]: W1215 12:40:21.833173 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde640a25_d092_4c73_bb2f_c7a21993261e.slice/crio-073ac7817c37ab63faa1b3e17c81e375e655ad106a1f5542cb09228f2404384d WatchSource:0}: Error finding container 073ac7817c37ab63faa1b3e17c81e375e655ad106a1f5542cb09228f2404384d: Status 404 returned error can't find the container with id 073ac7817c37ab63faa1b3e17c81e375e655ad106a1f5542cb09228f2404384d Dec 15 12:40:21 crc kubenswrapper[4719]: I1215 12:40:21.922805 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de640a25-d092-4c73-bb2f-c7a21993261e","Type":"ContainerStarted","Data":"073ac7817c37ab63faa1b3e17c81e375e655ad106a1f5542cb09228f2404384d"} Dec 15 12:40:27 crc kubenswrapper[4719]: I1215 12:40:27.389917 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.533589 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sf2kz"] Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.535786 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.549394 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sf2kz"] Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.657729 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-utilities\") pod \"certified-operators-sf2kz\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.657796 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlnn4\" (UniqueName: \"kubernetes.io/projected/cfa6641e-ffb8-483d-b153-15cd61500cfd-kube-api-access-jlnn4\") pod \"certified-operators-sf2kz\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.657818 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-catalog-content\") pod \"certified-operators-sf2kz\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.759962 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-utilities\") pod \"certified-operators-sf2kz\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.760042 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlnn4\" (UniqueName: \"kubernetes.io/projected/cfa6641e-ffb8-483d-b153-15cd61500cfd-kube-api-access-jlnn4\") pod \"certified-operators-sf2kz\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.760070 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-catalog-content\") pod \"certified-operators-sf2kz\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.760755 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-catalog-content\") pod \"certified-operators-sf2kz\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.761097 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-utilities\") pod \"certified-operators-sf2kz\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.778965 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlnn4\" (UniqueName: \"kubernetes.io/projected/cfa6641e-ffb8-483d-b153-15cd61500cfd-kube-api-access-jlnn4\") pod \"certified-operators-sf2kz\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:28 crc kubenswrapper[4719]: I1215 12:40:28.865833 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:29 crc kubenswrapper[4719]: I1215 12:40:29.017888 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de640a25-d092-4c73-bb2f-c7a21993261e","Type":"ContainerStarted","Data":"0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75"} Dec 15 12:40:29 crc kubenswrapper[4719]: I1215 12:40:29.035767 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x6vc4" event={"ID":"f1f092f5-6e7a-408d-8985-0a900417af4e","Type":"ContainerStarted","Data":"395f7c82e3a73edfeb1672d4ba7a438f9553821b569be491822069549733d9c0"} Dec 15 12:40:29 crc kubenswrapper[4719]: I1215 12:40:29.431304 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-x6vc4" podStartSLOduration=2.826486124 podStartE2EDuration="12.431287042s" podCreationTimestamp="2025-12-15 12:40:17 +0000 UTC" firstStartedPulling="2025-12-15 12:40:18.662831767 +0000 UTC m=+1379.605124797" lastFinishedPulling="2025-12-15 12:40:28.267632665 +0000 UTC m=+1389.209925715" observedRunningTime="2025-12-15 12:40:29.06117072 +0000 UTC m=+1390.003463750" watchObservedRunningTime="2025-12-15 12:40:29.431287042 +0000 UTC m=+1390.373580072" Dec 15 12:40:29 crc kubenswrapper[4719]: I1215 12:40:29.432209 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sf2kz"] Dec 15 12:40:30 crc kubenswrapper[4719]: I1215 12:40:30.049263 4719 generic.go:334] "Generic (PLEG): container finished" podID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerID="b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e" exitCode=0 Dec 15 12:40:30 crc kubenswrapper[4719]: I1215 12:40:30.049483 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sf2kz" event={"ID":"cfa6641e-ffb8-483d-b153-15cd61500cfd","Type":"ContainerDied","Data":"b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e"} Dec 15 12:40:30 crc kubenswrapper[4719]: I1215 12:40:30.050019 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sf2kz" event={"ID":"cfa6641e-ffb8-483d-b153-15cd61500cfd","Type":"ContainerStarted","Data":"e537dd82063bbbb389c8562509b0d3cb47aea563e6935c8fd35c7e486256f210"} Dec 15 12:40:31 crc kubenswrapper[4719]: I1215 12:40:31.066204 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de640a25-d092-4c73-bb2f-c7a21993261e","Type":"ContainerStarted","Data":"2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8"} Dec 15 12:40:31 crc kubenswrapper[4719]: I1215 12:40:31.066563 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de640a25-d092-4c73-bb2f-c7a21993261e","Type":"ContainerStarted","Data":"d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54"} Dec 15 12:40:31 crc kubenswrapper[4719]: I1215 12:40:31.521251 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:40:32 crc kubenswrapper[4719]: I1215 12:40:32.082540 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sf2kz" event={"ID":"cfa6641e-ffb8-483d-b153-15cd61500cfd","Type":"ContainerStarted","Data":"6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb"} Dec 15 12:40:33 crc kubenswrapper[4719]: I1215 12:40:33.093298 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de640a25-d092-4c73-bb2f-c7a21993261e","Type":"ContainerStarted","Data":"1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f"} Dec 15 12:40:33 crc kubenswrapper[4719]: I1215 12:40:33.096145 4719 generic.go:334] "Generic (PLEG): container finished" podID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerID="6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb" exitCode=0 Dec 15 12:40:33 crc kubenswrapper[4719]: I1215 12:40:33.096185 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sf2kz" event={"ID":"cfa6641e-ffb8-483d-b153-15cd61500cfd","Type":"ContainerDied","Data":"6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb"} Dec 15 12:40:33 crc kubenswrapper[4719]: I1215 12:40:33.972489 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6bc5fb48-jqf6b" Dec 15 12:40:34 crc kubenswrapper[4719]: I1215 12:40:34.053514 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c6bcf58b6-vx99l"] Dec 15 12:40:34 crc kubenswrapper[4719]: I1215 12:40:34.054128 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon-log" containerID="cri-o://c255291bba1cc5f6daef1060c061c10c9363cee215f9bd734f39e210ffbe638b" gracePeriod=30 Dec 15 12:40:34 crc kubenswrapper[4719]: I1215 12:40:34.054590 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c6bcf58b6-vx99l" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" containerID="cri-o://e1016fcc51f3a06758720611065b641214cb05d2109d1fc991685adda2a4994a" gracePeriod=30 Dec 15 12:40:34 crc kubenswrapper[4719]: I1215 12:40:34.142193 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sf2kz" event={"ID":"cfa6641e-ffb8-483d-b153-15cd61500cfd","Type":"ContainerStarted","Data":"7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6"} Dec 15 12:40:34 crc kubenswrapper[4719]: I1215 12:40:34.142445 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 15 12:40:34 crc kubenswrapper[4719]: I1215 12:40:34.164540 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.666455273 podStartE2EDuration="14.164516462s" podCreationTimestamp="2025-12-15 12:40:20 +0000 UTC" firstStartedPulling="2025-12-15 12:40:21.835130528 +0000 UTC m=+1382.777423558" lastFinishedPulling="2025-12-15 12:40:32.333191697 +0000 UTC m=+1393.275484747" observedRunningTime="2025-12-15 12:40:34.161961082 +0000 UTC m=+1395.104254112" watchObservedRunningTime="2025-12-15 12:40:34.164516462 +0000 UTC m=+1395.106809492" Dec 15 12:40:34 crc kubenswrapper[4719]: I1215 12:40:34.195155 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sf2kz" podStartSLOduration=2.595285889 podStartE2EDuration="6.195134162s" podCreationTimestamp="2025-12-15 12:40:28 +0000 UTC" firstStartedPulling="2025-12-15 12:40:30.050980697 +0000 UTC m=+1390.993273727" lastFinishedPulling="2025-12-15 12:40:33.65082897 +0000 UTC m=+1394.593122000" observedRunningTime="2025-12-15 12:40:34.177321023 +0000 UTC m=+1395.119614073" watchObservedRunningTime="2025-12-15 12:40:34.195134162 +0000 UTC m=+1395.137427192" Dec 15 12:40:35 crc kubenswrapper[4719]: I1215 12:40:35.167667 4719 generic.go:334] "Generic (PLEG): container finished" podID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerID="e1016fcc51f3a06758720611065b641214cb05d2109d1fc991685adda2a4994a" exitCode=0 Dec 15 12:40:35 crc kubenswrapper[4719]: I1215 12:40:35.167761 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6bcf58b6-vx99l" event={"ID":"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e","Type":"ContainerDied","Data":"e1016fcc51f3a06758720611065b641214cb05d2109d1fc991685adda2a4994a"} Dec 15 12:40:35 crc kubenswrapper[4719]: I1215 12:40:35.167832 4719 scope.go:117] "RemoveContainer" containerID="e526058563f282b4d65e64a01944ac0f829154bf6a70328411350711beca2af4" Dec 15 12:40:38 crc kubenswrapper[4719]: I1215 12:40:38.866193 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:38 crc kubenswrapper[4719]: I1215 12:40:38.867999 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:38 crc kubenswrapper[4719]: I1215 12:40:38.909284 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:39 crc kubenswrapper[4719]: I1215 12:40:39.245320 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:39 crc kubenswrapper[4719]: I1215 12:40:39.298736 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sf2kz"] Dec 15 12:40:41 crc kubenswrapper[4719]: I1215 12:40:41.214928 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sf2kz" podUID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerName="registry-server" containerID="cri-o://7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6" gracePeriod=2 Dec 15 12:40:41 crc kubenswrapper[4719]: E1215 12:40:41.469128 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfa6641e_ffb8_483d_b153_15cd61500cfd.slice/crio-conmon-7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfa6641e_ffb8_483d_b153_15cd61500cfd.slice/crio-7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6.scope\": RecentStats: unable to find data in memory cache]" Dec 15 12:40:41 crc kubenswrapper[4719]: I1215 12:40:41.726597 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:41 crc kubenswrapper[4719]: I1215 12:40:41.922961 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-catalog-content\") pod \"cfa6641e-ffb8-483d-b153-15cd61500cfd\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " Dec 15 12:40:41 crc kubenswrapper[4719]: I1215 12:40:41.923065 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlnn4\" (UniqueName: \"kubernetes.io/projected/cfa6641e-ffb8-483d-b153-15cd61500cfd-kube-api-access-jlnn4\") pod \"cfa6641e-ffb8-483d-b153-15cd61500cfd\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " Dec 15 12:40:41 crc kubenswrapper[4719]: I1215 12:40:41.923223 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-utilities\") pod \"cfa6641e-ffb8-483d-b153-15cd61500cfd\" (UID: \"cfa6641e-ffb8-483d-b153-15cd61500cfd\") " Dec 15 12:40:41 crc kubenswrapper[4719]: I1215 12:40:41.924096 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-utilities" (OuterVolumeSpecName: "utilities") pod "cfa6641e-ffb8-483d-b153-15cd61500cfd" (UID: "cfa6641e-ffb8-483d-b153-15cd61500cfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:40:41 crc kubenswrapper[4719]: I1215 12:40:41.929990 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfa6641e-ffb8-483d-b153-15cd61500cfd-kube-api-access-jlnn4" (OuterVolumeSpecName: "kube-api-access-jlnn4") pod "cfa6641e-ffb8-483d-b153-15cd61500cfd" (UID: "cfa6641e-ffb8-483d-b153-15cd61500cfd"). InnerVolumeSpecName "kube-api-access-jlnn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:40:41 crc kubenswrapper[4719]: I1215 12:40:41.981671 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfa6641e-ffb8-483d-b153-15cd61500cfd" (UID: "cfa6641e-ffb8-483d-b153-15cd61500cfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.025990 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlnn4\" (UniqueName: \"kubernetes.io/projected/cfa6641e-ffb8-483d-b153-15cd61500cfd-kube-api-access-jlnn4\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.026028 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.026045 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa6641e-ffb8-483d-b153-15cd61500cfd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.224817 4719 generic.go:334] "Generic (PLEG): container finished" podID="f1f092f5-6e7a-408d-8985-0a900417af4e" containerID="395f7c82e3a73edfeb1672d4ba7a438f9553821b569be491822069549733d9c0" exitCode=0 Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.224892 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x6vc4" event={"ID":"f1f092f5-6e7a-408d-8985-0a900417af4e","Type":"ContainerDied","Data":"395f7c82e3a73edfeb1672d4ba7a438f9553821b569be491822069549733d9c0"} Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.228681 4719 generic.go:334] "Generic (PLEG): container finished" podID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerID="7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6" exitCode=0 Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.228744 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sf2kz" event={"ID":"cfa6641e-ffb8-483d-b153-15cd61500cfd","Type":"ContainerDied","Data":"7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6"} Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.228770 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sf2kz" event={"ID":"cfa6641e-ffb8-483d-b153-15cd61500cfd","Type":"ContainerDied","Data":"e537dd82063bbbb389c8562509b0d3cb47aea563e6935c8fd35c7e486256f210"} Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.228819 4719 scope.go:117] "RemoveContainer" containerID="7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.229007 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sf2kz" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.255377 4719 scope.go:117] "RemoveContainer" containerID="6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.286534 4719 scope.go:117] "RemoveContainer" containerID="b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.292358 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sf2kz"] Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.299625 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sf2kz"] Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.318756 4719 scope.go:117] "RemoveContainer" containerID="7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6" Dec 15 12:40:42 crc kubenswrapper[4719]: E1215 12:40:42.319318 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6\": container with ID starting with 7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6 not found: ID does not exist" containerID="7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.319357 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6"} err="failed to get container status \"7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6\": rpc error: code = NotFound desc = could not find container \"7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6\": container with ID starting with 7dc454dbd650395797ac6c09c245fb32eab19e08bedf1b574a6ba243eec999d6 not found: ID does not exist" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.319383 4719 scope.go:117] "RemoveContainer" containerID="6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb" Dec 15 12:40:42 crc kubenswrapper[4719]: E1215 12:40:42.319671 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb\": container with ID starting with 6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb not found: ID does not exist" containerID="6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.319697 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb"} err="failed to get container status \"6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb\": rpc error: code = NotFound desc = could not find container \"6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb\": container with ID starting with 6f6878b3e1b6d5d43ab90efe1f01fa344e435bb465c07d92e137084288f21edb not found: ID does not exist" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.319713 4719 scope.go:117] "RemoveContainer" containerID="b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e" Dec 15 12:40:42 crc kubenswrapper[4719]: E1215 12:40:42.320073 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e\": container with ID starting with b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e not found: ID does not exist" containerID="b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e" Dec 15 12:40:42 crc kubenswrapper[4719]: I1215 12:40:42.320101 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e"} err="failed to get container status \"b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e\": rpc error: code = NotFound desc = could not find container \"b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e\": container with ID starting with b6b0edc94646b37ea44f7a5d284a4e02d214c80592913d30daf3e410e8a3d05e not found: ID does not exist" Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.493802 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfa6641e-ffb8-483d-b153-15cd61500cfd" path="/var/lib/kubelet/pods/cfa6641e-ffb8-483d-b153-15cd61500cfd/volumes" Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.577111 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.763061 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-scripts\") pod \"f1f092f5-6e7a-408d-8985-0a900417af4e\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.763124 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-config-data\") pod \"f1f092f5-6e7a-408d-8985-0a900417af4e\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.763227 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vcqr\" (UniqueName: \"kubernetes.io/projected/f1f092f5-6e7a-408d-8985-0a900417af4e-kube-api-access-7vcqr\") pod \"f1f092f5-6e7a-408d-8985-0a900417af4e\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.763362 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-combined-ca-bundle\") pod \"f1f092f5-6e7a-408d-8985-0a900417af4e\" (UID: \"f1f092f5-6e7a-408d-8985-0a900417af4e\") " Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.768252 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-scripts" (OuterVolumeSpecName: "scripts") pod "f1f092f5-6e7a-408d-8985-0a900417af4e" (UID: "f1f092f5-6e7a-408d-8985-0a900417af4e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.768258 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1f092f5-6e7a-408d-8985-0a900417af4e-kube-api-access-7vcqr" (OuterVolumeSpecName: "kube-api-access-7vcqr") pod "f1f092f5-6e7a-408d-8985-0a900417af4e" (UID: "f1f092f5-6e7a-408d-8985-0a900417af4e"). InnerVolumeSpecName "kube-api-access-7vcqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.787555 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1f092f5-6e7a-408d-8985-0a900417af4e" (UID: "f1f092f5-6e7a-408d-8985-0a900417af4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.792649 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-config-data" (OuterVolumeSpecName: "config-data") pod "f1f092f5-6e7a-408d-8985-0a900417af4e" (UID: "f1f092f5-6e7a-408d-8985-0a900417af4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.865916 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.865969 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vcqr\" (UniqueName: \"kubernetes.io/projected/f1f092f5-6e7a-408d-8985-0a900417af4e-kube-api-access-7vcqr\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.865982 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:43 crc kubenswrapper[4719]: I1215 12:40:43.865990 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1f092f5-6e7a-408d-8985-0a900417af4e-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.252436 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x6vc4" event={"ID":"f1f092f5-6e7a-408d-8985-0a900417af4e","Type":"ContainerDied","Data":"d70a8f04742d716a3313cc22d23a740a92bba16013420a7fc2d1283890491f53"} Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.252488 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d70a8f04742d716a3313cc22d23a740a92bba16013420a7fc2d1283890491f53" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.252933 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x6vc4" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.365598 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 15 12:40:44 crc kubenswrapper[4719]: E1215 12:40:44.366284 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1f092f5-6e7a-408d-8985-0a900417af4e" containerName="nova-cell0-conductor-db-sync" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.366301 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1f092f5-6e7a-408d-8985-0a900417af4e" containerName="nova-cell0-conductor-db-sync" Dec 15 12:40:44 crc kubenswrapper[4719]: E1215 12:40:44.366312 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerName="extract-utilities" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.366318 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerName="extract-utilities" Dec 15 12:40:44 crc kubenswrapper[4719]: E1215 12:40:44.366354 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerName="registry-server" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.366364 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerName="registry-server" Dec 15 12:40:44 crc kubenswrapper[4719]: E1215 12:40:44.366381 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerName="extract-content" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.366389 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerName="extract-content" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.366564 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1f092f5-6e7a-408d-8985-0a900417af4e" containerName="nova-cell0-conductor-db-sync" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.366580 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfa6641e-ffb8-483d-b153-15cd61500cfd" containerName="registry-server" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.367255 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.374635 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gj44p" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.374697 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.384511 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.475042 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2da200b-26c0-4242-a443-fa579e6da8da-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d2da200b-26c0-4242-a443-fa579e6da8da\") " pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.475209 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2da200b-26c0-4242-a443-fa579e6da8da-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d2da200b-26c0-4242-a443-fa579e6da8da\") " pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.475242 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwn6w\" (UniqueName: \"kubernetes.io/projected/d2da200b-26c0-4242-a443-fa579e6da8da-kube-api-access-zwn6w\") pod \"nova-cell0-conductor-0\" (UID: \"d2da200b-26c0-4242-a443-fa579e6da8da\") " pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.577384 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2da200b-26c0-4242-a443-fa579e6da8da-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d2da200b-26c0-4242-a443-fa579e6da8da\") " pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.577428 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwn6w\" (UniqueName: \"kubernetes.io/projected/d2da200b-26c0-4242-a443-fa579e6da8da-kube-api-access-zwn6w\") pod \"nova-cell0-conductor-0\" (UID: \"d2da200b-26c0-4242-a443-fa579e6da8da\") " pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.577536 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2da200b-26c0-4242-a443-fa579e6da8da-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d2da200b-26c0-4242-a443-fa579e6da8da\") " pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.582645 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2da200b-26c0-4242-a443-fa579e6da8da-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d2da200b-26c0-4242-a443-fa579e6da8da\") " pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.589496 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2da200b-26c0-4242-a443-fa579e6da8da-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d2da200b-26c0-4242-a443-fa579e6da8da\") " pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.598795 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwn6w\" (UniqueName: \"kubernetes.io/projected/d2da200b-26c0-4242-a443-fa579e6da8da-kube-api-access-zwn6w\") pod \"nova-cell0-conductor-0\" (UID: \"d2da200b-26c0-4242-a443-fa579e6da8da\") " pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:44 crc kubenswrapper[4719]: I1215 12:40:44.691932 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:45 crc kubenswrapper[4719]: I1215 12:40:45.144460 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 15 12:40:45 crc kubenswrapper[4719]: W1215 12:40:45.150029 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2da200b_26c0_4242_a443_fa579e6da8da.slice/crio-bfbecbcacd73cdc6832f3cbf271054fde1e0a6ae3a36a8ba23d4dfdf978e44e6 WatchSource:0}: Error finding container bfbecbcacd73cdc6832f3cbf271054fde1e0a6ae3a36a8ba23d4dfdf978e44e6: Status 404 returned error can't find the container with id bfbecbcacd73cdc6832f3cbf271054fde1e0a6ae3a36a8ba23d4dfdf978e44e6 Dec 15 12:40:45 crc kubenswrapper[4719]: I1215 12:40:45.261003 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d2da200b-26c0-4242-a443-fa579e6da8da","Type":"ContainerStarted","Data":"bfbecbcacd73cdc6832f3cbf271054fde1e0a6ae3a36a8ba23d4dfdf978e44e6"} Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.270544 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d2da200b-26c0-4242-a443-fa579e6da8da","Type":"ContainerStarted","Data":"85d6b721e9f7548d3bad06306dcc7bfd62475ce08ba18e49d90c180202a2251b"} Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.270928 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.295160 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.295138416 podStartE2EDuration="2.295138416s" podCreationTimestamp="2025-12-15 12:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:40:46.293083782 +0000 UTC m=+1407.235376802" watchObservedRunningTime="2025-12-15 12:40:46.295138416 +0000 UTC m=+1407.237431446" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.703746 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vsgzc"] Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.706024 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.742018 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vsgzc"] Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.817210 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7nml\" (UniqueName: \"kubernetes.io/projected/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-kube-api-access-r7nml\") pod \"redhat-operators-vsgzc\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.817302 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-catalog-content\") pod \"redhat-operators-vsgzc\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.817511 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-utilities\") pod \"redhat-operators-vsgzc\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.919764 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7nml\" (UniqueName: \"kubernetes.io/projected/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-kube-api-access-r7nml\") pod \"redhat-operators-vsgzc\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.919846 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-catalog-content\") pod \"redhat-operators-vsgzc\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.919881 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-utilities\") pod \"redhat-operators-vsgzc\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.920756 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-catalog-content\") pod \"redhat-operators-vsgzc\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.920764 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-utilities\") pod \"redhat-operators-vsgzc\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:46 crc kubenswrapper[4719]: I1215 12:40:46.937825 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7nml\" (UniqueName: \"kubernetes.io/projected/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-kube-api-access-r7nml\") pod \"redhat-operators-vsgzc\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:47 crc kubenswrapper[4719]: I1215 12:40:47.042677 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:47 crc kubenswrapper[4719]: I1215 12:40:47.509100 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vsgzc"] Dec 15 12:40:47 crc kubenswrapper[4719]: W1215 12:40:47.509936 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ecc97bf_b6b8_4a59_a449_ca8b19c6f658.slice/crio-2dc9f622c00d9895fab4f748d96d7374598655704457cbc4d3756448b646a3ec WatchSource:0}: Error finding container 2dc9f622c00d9895fab4f748d96d7374598655704457cbc4d3756448b646a3ec: Status 404 returned error can't find the container with id 2dc9f622c00d9895fab4f748d96d7374598655704457cbc4d3756448b646a3ec Dec 15 12:40:48 crc kubenswrapper[4719]: I1215 12:40:48.290543 4719 generic.go:334] "Generic (PLEG): container finished" podID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerID="875449d5418281b692ac2c63f1df67e3142449e6f43d331bc5f7112a5f9fa9c2" exitCode=0 Dec 15 12:40:48 crc kubenswrapper[4719]: I1215 12:40:48.290619 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsgzc" event={"ID":"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658","Type":"ContainerDied","Data":"875449d5418281b692ac2c63f1df67e3142449e6f43d331bc5f7112a5f9fa9c2"} Dec 15 12:40:48 crc kubenswrapper[4719]: I1215 12:40:48.291162 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsgzc" event={"ID":"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658","Type":"ContainerStarted","Data":"2dc9f622c00d9895fab4f748d96d7374598655704457cbc4d3756448b646a3ec"} Dec 15 12:40:50 crc kubenswrapper[4719]: I1215 12:40:50.308686 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsgzc" event={"ID":"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658","Type":"ContainerStarted","Data":"32671fc64276bddfb76780a433ff0be436c090048a677a3f07ccd3cdc3f7ab23"} Dec 15 12:40:51 crc kubenswrapper[4719]: I1215 12:40:51.333573 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 15 12:40:53 crc kubenswrapper[4719]: I1215 12:40:53.336122 4719 generic.go:334] "Generic (PLEG): container finished" podID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerID="32671fc64276bddfb76780a433ff0be436c090048a677a3f07ccd3cdc3f7ab23" exitCode=0 Dec 15 12:40:53 crc kubenswrapper[4719]: I1215 12:40:53.336595 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsgzc" event={"ID":"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658","Type":"ContainerDied","Data":"32671fc64276bddfb76780a433ff0be436c090048a677a3f07ccd3cdc3f7ab23"} Dec 15 12:40:54 crc kubenswrapper[4719]: I1215 12:40:54.719117 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.489947 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vzsnw"] Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.491020 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.505079 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vzsnw"] Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.509644 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.510411 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.696973 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj4z6\" (UniqueName: \"kubernetes.io/projected/30929e15-1877-4485-9366-6396d32df2ae-kube-api-access-fj4z6\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.697058 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-scripts\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.697265 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.697369 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-config-data\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.757187 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.770504 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.782313 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.789390 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.791069 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.801292 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.802879 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-config-data\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.802950 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj4z6\" (UniqueName: \"kubernetes.io/projected/30929e15-1877-4485-9366-6396d32df2ae-kube-api-access-fj4z6\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.802979 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-scripts\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.803066 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.814157 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-config-data\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.832806 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.859387 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-scripts\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.861247 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.870663 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.889969 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj4z6\" (UniqueName: \"kubernetes.io/projected/30929e15-1877-4485-9366-6396d32df2ae-kube-api-access-fj4z6\") pod \"nova-cell0-cell-mapping-vzsnw\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.894995 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.896267 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.906935 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-logs\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.907022 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-config-data\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.907074 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhgsp\" (UniqueName: \"kubernetes.io/projected/5c62a0a4-152d-4953-a2c3-ab9343c208ba-kube-api-access-bhgsp\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.907098 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzjlx\" (UniqueName: \"kubernetes.io/projected/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-kube-api-access-jzjlx\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.907118 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.907145 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.907161 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c62a0a4-152d-4953-a2c3-ab9343c208ba-logs\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.907183 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-config-data\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.910003 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 15 12:40:55 crc kubenswrapper[4719]: I1215 12:40:55.933041 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.009407 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-jd2vn"] Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.010923 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013104 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-config-data\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013140 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013160 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013203 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhgsp\" (UniqueName: \"kubernetes.io/projected/5c62a0a4-152d-4953-a2c3-ab9343c208ba-kube-api-access-bhgsp\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013223 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013239 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzjlx\" (UniqueName: \"kubernetes.io/projected/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-kube-api-access-jzjlx\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013257 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xg48\" (UniqueName: \"kubernetes.io/projected/8ee6b204-39a3-4189-8ca8-8210ed849824-kube-api-access-2xg48\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013274 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013297 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013312 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c62a0a4-152d-4953-a2c3-ab9343c208ba-logs\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013330 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-config-data\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013359 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-config\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013398 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-logs\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013413 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013433 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9c4f\" (UniqueName: \"kubernetes.io/projected/e8230fff-add6-4db5-b288-638524c1d197-kube-api-access-n9c4f\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013451 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.013476 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.021364 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.022256 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-logs\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.036202 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-jd2vn"] Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.037065 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c62a0a4-152d-4953-a2c3-ab9343c208ba-logs\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.038442 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.038956 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-config-data\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.058966 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-config-data\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.069681 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhgsp\" (UniqueName: \"kubernetes.io/projected/5c62a0a4-152d-4953-a2c3-ab9343c208ba-kube-api-access-bhgsp\") pod \"nova-api-0\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " pod="openstack/nova-api-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.077588 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.080566 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.082751 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.094123 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzjlx\" (UniqueName: \"kubernetes.io/projected/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-kube-api-access-jzjlx\") pod \"nova-metadata-0\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " pod="openstack/nova-metadata-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.115067 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.115368 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xg48\" (UniqueName: \"kubernetes.io/projected/8ee6b204-39a3-4189-8ca8-8210ed849824-kube-api-access-2xg48\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.115496 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79wlj\" (UniqueName: \"kubernetes.io/projected/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-kube-api-access-79wlj\") pod \"nova-scheduler-0\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.120788 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.123270 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-config\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.123416 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.123508 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-config-data\") pod \"nova-scheduler-0\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.123639 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.123762 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9c4f\" (UniqueName: \"kubernetes.io/projected/e8230fff-add6-4db5-b288-638524c1d197-kube-api-access-n9c4f\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.123899 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.124058 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.130380 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.130522 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.125883 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-config\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.126596 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.126675 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.129957 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.124642 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.125301 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.154049 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.154832 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.162213 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9c4f\" (UniqueName: \"kubernetes.io/projected/e8230fff-add6-4db5-b288-638524c1d197-kube-api-access-n9c4f\") pod \"dnsmasq-dns-845d6d6f59-jd2vn\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.172156 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xg48\" (UniqueName: \"kubernetes.io/projected/8ee6b204-39a3-4189-8ca8-8210ed849824-kube-api-access-2xg48\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.173410 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.233224 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79wlj\" (UniqueName: \"kubernetes.io/projected/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-kube-api-access-79wlj\") pod \"nova-scheduler-0\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.233523 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.233547 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-config-data\") pod \"nova-scheduler-0\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.233318 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.244401 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.244926 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-config-data\") pod \"nova-scheduler-0\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.261237 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79wlj\" (UniqueName: \"kubernetes.io/projected/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-kube-api-access-79wlj\") pod \"nova-scheduler-0\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.266406 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.304750 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.382126 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsgzc" event={"ID":"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658","Type":"ContainerStarted","Data":"7de42768b6b795add51e739ccd2eec25ac3ad6203ca2dd4fbf459c81a96feea9"} Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.407722 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vsgzc" podStartSLOduration=3.034141225 podStartE2EDuration="10.407705151s" podCreationTimestamp="2025-12-15 12:40:46 +0000 UTC" firstStartedPulling="2025-12-15 12:40:48.293062444 +0000 UTC m=+1409.235355474" lastFinishedPulling="2025-12-15 12:40:55.66662637 +0000 UTC m=+1416.608919400" observedRunningTime="2025-12-15 12:40:56.401084994 +0000 UTC m=+1417.343378024" watchObservedRunningTime="2025-12-15 12:40:56.407705151 +0000 UTC m=+1417.349998181" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.481656 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 15 12:40:56 crc kubenswrapper[4719]: I1215 12:40:56.822957 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vzsnw"] Dec 15 12:40:56 crc kubenswrapper[4719]: W1215 12:40:56.830734 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30929e15_1877_4485_9366_6396d32df2ae.slice/crio-e227a2c07eba898614ec6e8399d0b16edf2179f2e52f9dea187372a134123143 WatchSource:0}: Error finding container e227a2c07eba898614ec6e8399d0b16edf2179f2e52f9dea187372a134123143: Status 404 returned error can't find the container with id e227a2c07eba898614ec6e8399d0b16edf2179f2e52f9dea187372a134123143 Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.021809 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.045408 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.045680 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.211377 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.225063 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.286771 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.302247 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-jd2vn"] Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.400763 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35b456e0-f0a8-40d3-a8fe-77b9880cecc4","Type":"ContainerStarted","Data":"8b9a6613ec8c9c9de5719e1ea9f3303b66652d161ff6239b9904a18b8c04e6d1"} Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.423818 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" event={"ID":"e8230fff-add6-4db5-b288-638524c1d197","Type":"ContainerStarted","Data":"1b4a54ebb2c4089f6df324830b4820cd9b6048ef58d22a8bfb670bb4021f7453"} Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.449466 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5c62a0a4-152d-4953-a2c3-ab9343c208ba","Type":"ContainerStarted","Data":"2e903192991901c1042ad1e32b98ad36af2453cf54f87b6fbbabc14b48c71e6c"} Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.449519 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8ee6b204-39a3-4189-8ca8-8210ed849824","Type":"ContainerStarted","Data":"0ee4c5fd432ab8a452541cb75785d19b7698682e78927cee127533842734bffd"} Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.449535 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"40707f09-eaf4-4dfc-bd8c-e4d1336c327b","Type":"ContainerStarted","Data":"a7c0f25f95f8b8d9fbb100f1242cebd2b9ce58de7e8688fb50e8c7f37342c77f"} Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.457957 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vzsnw" event={"ID":"30929e15-1877-4485-9366-6396d32df2ae","Type":"ContainerStarted","Data":"166b49effe874c8c992db325f19173be0def7b459985ad660ca508d4805b4f8f"} Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.458048 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vzsnw" event={"ID":"30929e15-1877-4485-9366-6396d32df2ae","Type":"ContainerStarted","Data":"e227a2c07eba898614ec6e8399d0b16edf2179f2e52f9dea187372a134123143"} Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.472685 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5p65f"] Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.482510 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.490003 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.490083 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.499043 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vzsnw" podStartSLOduration=2.49902415 podStartE2EDuration="2.49902415s" podCreationTimestamp="2025-12-15 12:40:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:40:57.483446023 +0000 UTC m=+1418.425739063" watchObservedRunningTime="2025-12-15 12:40:57.49902415 +0000 UTC m=+1418.441317190" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.509095 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5p65f"] Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.564785 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-config-data\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.566077 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.566288 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-scripts\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.566545 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4z24\" (UniqueName: \"kubernetes.io/projected/1c9e9ff2-3aa8-4640-979a-be62e16eca47-kube-api-access-b4z24\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.668522 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4z24\" (UniqueName: \"kubernetes.io/projected/1c9e9ff2-3aa8-4640-979a-be62e16eca47-kube-api-access-b4z24\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.668904 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-config-data\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.669026 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.669552 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-scripts\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.673832 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-scripts\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.675147 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-config-data\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.678457 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.689551 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4z24\" (UniqueName: \"kubernetes.io/projected/1c9e9ff2-3aa8-4640-979a-be62e16eca47-kube-api-access-b4z24\") pod \"nova-cell1-conductor-db-sync-5p65f\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:57 crc kubenswrapper[4719]: I1215 12:40:57.827296 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:40:58 crc kubenswrapper[4719]: I1215 12:40:58.134070 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vsgzc" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="registry-server" probeResult="failure" output=< Dec 15 12:40:58 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 12:40:58 crc kubenswrapper[4719]: > Dec 15 12:40:58 crc kubenswrapper[4719]: I1215 12:40:58.414653 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5p65f"] Dec 15 12:40:58 crc kubenswrapper[4719]: W1215 12:40:58.461210 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c9e9ff2_3aa8_4640_979a_be62e16eca47.slice/crio-f6eac6e18dcf460771584cd39cf1d0c794cc03adc1606e10a6868dc61531b517 WatchSource:0}: Error finding container f6eac6e18dcf460771584cd39cf1d0c794cc03adc1606e10a6868dc61531b517: Status 404 returned error can't find the container with id f6eac6e18dcf460771584cd39cf1d0c794cc03adc1606e10a6868dc61531b517 Dec 15 12:40:58 crc kubenswrapper[4719]: I1215 12:40:58.480930 4719 generic.go:334] "Generic (PLEG): container finished" podID="e8230fff-add6-4db5-b288-638524c1d197" containerID="e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10" exitCode=0 Dec 15 12:40:58 crc kubenswrapper[4719]: I1215 12:40:58.480996 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" event={"ID":"e8230fff-add6-4db5-b288-638524c1d197","Type":"ContainerDied","Data":"e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10"} Dec 15 12:40:59 crc kubenswrapper[4719]: I1215 12:40:59.523202 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" event={"ID":"e8230fff-add6-4db5-b288-638524c1d197","Type":"ContainerStarted","Data":"479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe"} Dec 15 12:40:59 crc kubenswrapper[4719]: I1215 12:40:59.524451 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:40:59 crc kubenswrapper[4719]: I1215 12:40:59.537666 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5p65f" event={"ID":"1c9e9ff2-3aa8-4640-979a-be62e16eca47","Type":"ContainerStarted","Data":"4504d72e3231d8806da3f3047b136c2e704fefb6a67479e700f38fc11bc789df"} Dec 15 12:40:59 crc kubenswrapper[4719]: I1215 12:40:59.537727 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5p65f" event={"ID":"1c9e9ff2-3aa8-4640-979a-be62e16eca47","Type":"ContainerStarted","Data":"f6eac6e18dcf460771584cd39cf1d0c794cc03adc1606e10a6868dc61531b517"} Dec 15 12:40:59 crc kubenswrapper[4719]: I1215 12:40:59.639438 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" podStartSLOduration=4.639415434 podStartE2EDuration="4.639415434s" podCreationTimestamp="2025-12-15 12:40:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:40:59.627546883 +0000 UTC m=+1420.569839913" watchObservedRunningTime="2025-12-15 12:40:59.639415434 +0000 UTC m=+1420.581708464" Dec 15 12:40:59 crc kubenswrapper[4719]: I1215 12:40:59.651014 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-5p65f" podStartSLOduration=2.650992487 podStartE2EDuration="2.650992487s" podCreationTimestamp="2025-12-15 12:40:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:40:59.646746284 +0000 UTC m=+1420.589039314" watchObservedRunningTime="2025-12-15 12:40:59.650992487 +0000 UTC m=+1420.593285527" Dec 15 12:41:00 crc kubenswrapper[4719]: I1215 12:41:00.456165 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:00 crc kubenswrapper[4719]: I1215 12:41:00.468222 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.615373 4719 generic.go:334] "Generic (PLEG): container finished" podID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerID="c255291bba1cc5f6daef1060c061c10c9363cee215f9bd734f39e210ffbe638b" exitCode=137 Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.615468 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6bcf58b6-vx99l" event={"ID":"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e","Type":"ContainerDied","Data":"c255291bba1cc5f6daef1060c061c10c9363cee215f9bd734f39e210ffbe638b"} Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.694316 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.832458 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-config-data\") pod \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.832736 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-tls-certs\") pod \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.832914 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8b2b\" (UniqueName: \"kubernetes.io/projected/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-kube-api-access-b8b2b\") pod \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.833121 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-scripts\") pod \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.833245 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-combined-ca-bundle\") pod \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.833436 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-logs\") pod \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.833570 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-secret-key\") pod \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\" (UID: \"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e\") " Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.838470 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-logs" (OuterVolumeSpecName: "logs") pod "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" (UID: "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.841528 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" (UID: "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.841724 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-kube-api-access-b8b2b" (OuterVolumeSpecName: "kube-api-access-b8b2b") pod "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" (UID: "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e"). InnerVolumeSpecName "kube-api-access-b8b2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.867210 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-config-data" (OuterVolumeSpecName: "config-data") pod "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" (UID: "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.871608 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" (UID: "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.886527 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-scripts" (OuterVolumeSpecName: "scripts") pod "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" (UID: "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.914480 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" (UID: "ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.937364 4719 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.937667 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8b2b\" (UniqueName: \"kubernetes.io/projected/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-kube-api-access-b8b2b\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.937681 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.937695 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.937708 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.937718 4719 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:04 crc kubenswrapper[4719]: I1215 12:41:04.937728 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.637791 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5c62a0a4-152d-4953-a2c3-ab9343c208ba","Type":"ContainerStarted","Data":"873d279897789b83274cf74d5e7acd31b8cebc22b0e832ae5fc624a833e1c844"} Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.637843 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5c62a0a4-152d-4953-a2c3-ab9343c208ba","Type":"ContainerStarted","Data":"bd6c8d4b6435307bdf79d1759e4f8dfe321baf21ca202c9f61f0e5191456e363"} Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.642298 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8ee6b204-39a3-4189-8ca8-8210ed849824","Type":"ContainerStarted","Data":"e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e"} Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.642404 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="8ee6b204-39a3-4189-8ca8-8210ed849824" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e" gracePeriod=30 Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.644328 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"40707f09-eaf4-4dfc-bd8c-e4d1336c327b","Type":"ContainerStarted","Data":"d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d"} Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.649076 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c6bcf58b6-vx99l" event={"ID":"ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e","Type":"ContainerDied","Data":"228c4f698c30ff9eb50ccbb80912dae089a8469028a3ce638228519ede1ea32b"} Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.649126 4719 scope.go:117] "RemoveContainer" containerID="e1016fcc51f3a06758720611065b641214cb05d2109d1fc991685adda2a4994a" Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.649255 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c6bcf58b6-vx99l" Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.664532 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35b456e0-f0a8-40d3-a8fe-77b9880cecc4","Type":"ContainerStarted","Data":"e2f32b7d40fdda747ae373cddfc166978b76b9cadd477bd4bec8ba4aee3c72a1"} Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.664581 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35b456e0-f0a8-40d3-a8fe-77b9880cecc4","Type":"ContainerStarted","Data":"c5edcc8e7d711ec5ab47fed17ee5471a1929c0fc10ed620ab852ddf7f7566937"} Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.664927 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" containerName="nova-metadata-log" containerID="cri-o://c5edcc8e7d711ec5ab47fed17ee5471a1929c0fc10ed620ab852ddf7f7566937" gracePeriod=30 Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.664964 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" containerName="nova-metadata-metadata" containerID="cri-o://e2f32b7d40fdda747ae373cddfc166978b76b9cadd477bd4bec8ba4aee3c72a1" gracePeriod=30 Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.666233 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.379607732 podStartE2EDuration="10.666219032s" podCreationTimestamp="2025-12-15 12:40:55 +0000 UTC" firstStartedPulling="2025-12-15 12:40:57.051007117 +0000 UTC m=+1417.993300147" lastFinishedPulling="2025-12-15 12:41:04.337618417 +0000 UTC m=+1425.279911447" observedRunningTime="2025-12-15 12:41:05.660302206 +0000 UTC m=+1426.602595246" watchObservedRunningTime="2025-12-15 12:41:05.666219032 +0000 UTC m=+1426.608512082" Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.694975 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c6bcf58b6-vx99l"] Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.707193 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c6bcf58b6-vx99l"] Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.710609 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.844093825 podStartE2EDuration="10.710587036s" podCreationTimestamp="2025-12-15 12:40:55 +0000 UTC" firstStartedPulling="2025-12-15 12:40:57.235976295 +0000 UTC m=+1418.178269325" lastFinishedPulling="2025-12-15 12:41:04.102469506 +0000 UTC m=+1425.044762536" observedRunningTime="2025-12-15 12:41:05.705814567 +0000 UTC m=+1426.648107597" watchObservedRunningTime="2025-12-15 12:41:05.710587036 +0000 UTC m=+1426.652880066" Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.724448 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.882633714 podStartE2EDuration="10.724425351s" podCreationTimestamp="2025-12-15 12:40:55 +0000 UTC" firstStartedPulling="2025-12-15 12:40:57.261840276 +0000 UTC m=+1418.204133306" lastFinishedPulling="2025-12-15 12:41:04.103631913 +0000 UTC m=+1425.045924943" observedRunningTime="2025-12-15 12:41:05.721598602 +0000 UTC m=+1426.663891652" watchObservedRunningTime="2025-12-15 12:41:05.724425351 +0000 UTC m=+1426.666718381" Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.750764 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.94559531 podStartE2EDuration="10.750741888s" podCreationTimestamp="2025-12-15 12:40:55 +0000 UTC" firstStartedPulling="2025-12-15 12:40:57.303397418 +0000 UTC m=+1418.245690448" lastFinishedPulling="2025-12-15 12:41:04.108543996 +0000 UTC m=+1425.050837026" observedRunningTime="2025-12-15 12:41:05.7441145 +0000 UTC m=+1426.686407530" watchObservedRunningTime="2025-12-15 12:41:05.750741888 +0000 UTC m=+1426.693034918" Dec 15 12:41:05 crc kubenswrapper[4719]: I1215 12:41:05.946502 4719 scope.go:117] "RemoveContainer" containerID="c255291bba1cc5f6daef1060c061c10c9363cee215f9bd734f39e210ffbe638b" Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.176036 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.234639 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.234945 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.249051 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-b4mfs"] Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.249348 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" podUID="d68b9a30-55fe-4d32-8cba-9a12dcdc210c" containerName="dnsmasq-dns" containerID="cri-o://ff799566508d1613e80c3cd0de443e3cb223b7e9d5b48c1b857f9511ebaa7d30" gracePeriod=10 Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.266535 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.266598 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.307295 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.483590 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.483652 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.522270 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.675483 4719 generic.go:334] "Generic (PLEG): container finished" podID="d68b9a30-55fe-4d32-8cba-9a12dcdc210c" containerID="ff799566508d1613e80c3cd0de443e3cb223b7e9d5b48c1b857f9511ebaa7d30" exitCode=0 Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.675591 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" event={"ID":"d68b9a30-55fe-4d32-8cba-9a12dcdc210c","Type":"ContainerDied","Data":"ff799566508d1613e80c3cd0de443e3cb223b7e9d5b48c1b857f9511ebaa7d30"} Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.678132 4719 generic.go:334] "Generic (PLEG): container finished" podID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" containerID="e2f32b7d40fdda747ae373cddfc166978b76b9cadd477bd4bec8ba4aee3c72a1" exitCode=0 Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.678157 4719 generic.go:334] "Generic (PLEG): container finished" podID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" containerID="c5edcc8e7d711ec5ab47fed17ee5471a1929c0fc10ed620ab852ddf7f7566937" exitCode=143 Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.678205 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35b456e0-f0a8-40d3-a8fe-77b9880cecc4","Type":"ContainerDied","Data":"e2f32b7d40fdda747ae373cddfc166978b76b9cadd477bd4bec8ba4aee3c72a1"} Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.678225 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35b456e0-f0a8-40d3-a8fe-77b9880cecc4","Type":"ContainerDied","Data":"c5edcc8e7d711ec5ab47fed17ee5471a1929c0fc10ed620ab852ddf7f7566937"} Dec 15 12:41:06 crc kubenswrapper[4719]: I1215 12:41:06.723813 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 15 12:41:07 crc kubenswrapper[4719]: I1215 12:41:07.318120 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 15 12:41:07 crc kubenswrapper[4719]: I1215 12:41:07.318618 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 15 12:41:07 crc kubenswrapper[4719]: I1215 12:41:07.487956 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" path="/var/lib/kubelet/pods/ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e/volumes" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.112290 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vsgzc" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="registry-server" probeResult="failure" output=< Dec 15 12:41:08 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 12:41:08 crc kubenswrapper[4719]: > Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.126588 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.151044 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.231501 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-config\") pod \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.231587 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-nb\") pod \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.231633 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-swift-storage-0\") pod \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.231725 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvldx\" (UniqueName: \"kubernetes.io/projected/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-kube-api-access-qvldx\") pod \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.231768 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-sb\") pod \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.231846 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-svc\") pod \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\" (UID: \"d68b9a30-55fe-4d32-8cba-9a12dcdc210c\") " Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.246989 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-kube-api-access-qvldx" (OuterVolumeSpecName: "kube-api-access-qvldx") pod "d68b9a30-55fe-4d32-8cba-9a12dcdc210c" (UID: "d68b9a30-55fe-4d32-8cba-9a12dcdc210c"). InnerVolumeSpecName "kube-api-access-qvldx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.323299 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d68b9a30-55fe-4d32-8cba-9a12dcdc210c" (UID: "d68b9a30-55fe-4d32-8cba-9a12dcdc210c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.345935 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzjlx\" (UniqueName: \"kubernetes.io/projected/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-kube-api-access-jzjlx\") pod \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.345982 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-combined-ca-bundle\") pod \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.346091 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-logs\") pod \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.346131 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-config-data\") pod \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\" (UID: \"35b456e0-f0a8-40d3-a8fe-77b9880cecc4\") " Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.347154 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvldx\" (UniqueName: \"kubernetes.io/projected/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-kube-api-access-qvldx\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.347182 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.347378 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-config" (OuterVolumeSpecName: "config") pod "d68b9a30-55fe-4d32-8cba-9a12dcdc210c" (UID: "d68b9a30-55fe-4d32-8cba-9a12dcdc210c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.347652 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-logs" (OuterVolumeSpecName: "logs") pod "35b456e0-f0a8-40d3-a8fe-77b9880cecc4" (UID: "35b456e0-f0a8-40d3-a8fe-77b9880cecc4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.371771 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-kube-api-access-jzjlx" (OuterVolumeSpecName: "kube-api-access-jzjlx") pod "35b456e0-f0a8-40d3-a8fe-77b9880cecc4" (UID: "35b456e0-f0a8-40d3-a8fe-77b9880cecc4"). InnerVolumeSpecName "kube-api-access-jzjlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.378836 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d68b9a30-55fe-4d32-8cba-9a12dcdc210c" (UID: "d68b9a30-55fe-4d32-8cba-9a12dcdc210c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.383578 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35b456e0-f0a8-40d3-a8fe-77b9880cecc4" (UID: "35b456e0-f0a8-40d3-a8fe-77b9880cecc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.383673 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-config-data" (OuterVolumeSpecName: "config-data") pod "35b456e0-f0a8-40d3-a8fe-77b9880cecc4" (UID: "35b456e0-f0a8-40d3-a8fe-77b9880cecc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.401327 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d68b9a30-55fe-4d32-8cba-9a12dcdc210c" (UID: "d68b9a30-55fe-4d32-8cba-9a12dcdc210c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.426665 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d68b9a30-55fe-4d32-8cba-9a12dcdc210c" (UID: "d68b9a30-55fe-4d32-8cba-9a12dcdc210c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.449344 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.449623 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.449696 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.449763 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.449827 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.449935 4719 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d68b9a30-55fe-4d32-8cba-9a12dcdc210c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.450001 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzjlx\" (UniqueName: \"kubernetes.io/projected/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-kube-api-access-jzjlx\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.450061 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35b456e0-f0a8-40d3-a8fe-77b9880cecc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.714996 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" event={"ID":"d68b9a30-55fe-4d32-8cba-9a12dcdc210c","Type":"ContainerDied","Data":"ba157ab11cfe480035a4cf889637d80cb7113976b48c930cf382a7661c7c0852"} Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.715065 4719 scope.go:117] "RemoveContainer" containerID="ff799566508d1613e80c3cd0de443e3cb223b7e9d5b48c1b857f9511ebaa7d30" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.715202 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.737365 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.739202 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"35b456e0-f0a8-40d3-a8fe-77b9880cecc4","Type":"ContainerDied","Data":"8b9a6613ec8c9c9de5719e1ea9f3303b66652d161ff6239b9904a18b8c04e6d1"} Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.771251 4719 scope.go:117] "RemoveContainer" containerID="e0d877200e56dde9fde96c37bd8c6403c6359b013c28cd758f916771d3a39598" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.771435 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-b4mfs"] Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.777487 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-b4mfs"] Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.809576 4719 scope.go:117] "RemoveContainer" containerID="e2f32b7d40fdda747ae373cddfc166978b76b9cadd477bd4bec8ba4aee3c72a1" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.811900 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.819478 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.828837 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:08 crc kubenswrapper[4719]: E1215 12:41:08.829321 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829341 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" Dec 15 12:41:08 crc kubenswrapper[4719]: E1215 12:41:08.829363 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon-log" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829371 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon-log" Dec 15 12:41:08 crc kubenswrapper[4719]: E1215 12:41:08.829389 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d68b9a30-55fe-4d32-8cba-9a12dcdc210c" containerName="dnsmasq-dns" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829396 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d68b9a30-55fe-4d32-8cba-9a12dcdc210c" containerName="dnsmasq-dns" Dec 15 12:41:08 crc kubenswrapper[4719]: E1215 12:41:08.829406 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829411 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" Dec 15 12:41:08 crc kubenswrapper[4719]: E1215 12:41:08.829419 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d68b9a30-55fe-4d32-8cba-9a12dcdc210c" containerName="init" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829430 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d68b9a30-55fe-4d32-8cba-9a12dcdc210c" containerName="init" Dec 15 12:41:08 crc kubenswrapper[4719]: E1215 12:41:08.829459 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" containerName="nova-metadata-metadata" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829467 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" containerName="nova-metadata-metadata" Dec 15 12:41:08 crc kubenswrapper[4719]: E1215 12:41:08.829479 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" containerName="nova-metadata-log" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829484 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" containerName="nova-metadata-log" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829661 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829680 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829691 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829700 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" containerName="nova-metadata-metadata" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829708 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon-log" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829714 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" containerName="nova-metadata-log" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.829729 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d68b9a30-55fe-4d32-8cba-9a12dcdc210c" containerName="dnsmasq-dns" Dec 15 12:41:08 crc kubenswrapper[4719]: E1215 12:41:08.833059 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.833085 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2ae4f5-b3a6-4f26-b897-2dc0cb588f6e" containerName="horizon" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.834026 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.842570 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.843051 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.845881 4719 scope.go:117] "RemoveContainer" containerID="c5edcc8e7d711ec5ab47fed17ee5471a1929c0fc10ed620ab852ddf7f7566937" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.854189 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.959378 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.959487 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tkrx\" (UniqueName: \"kubernetes.io/projected/333a9883-5cc9-48cd-81cd-e1d87e91006c-kube-api-access-4tkrx\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.959558 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.959583 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/333a9883-5cc9-48cd-81cd-e1d87e91006c-logs\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:08 crc kubenswrapper[4719]: I1215 12:41:08.959623 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-config-data\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.061525 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.061560 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/333a9883-5cc9-48cd-81cd-e1d87e91006c-logs\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.061594 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-config-data\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.061683 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.061731 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tkrx\" (UniqueName: \"kubernetes.io/projected/333a9883-5cc9-48cd-81cd-e1d87e91006c-kube-api-access-4tkrx\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.062983 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/333a9883-5cc9-48cd-81cd-e1d87e91006c-logs\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.067670 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.068338 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.078320 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-config-data\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.083598 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tkrx\" (UniqueName: \"kubernetes.io/projected/333a9883-5cc9-48cd-81cd-e1d87e91006c-kube-api-access-4tkrx\") pod \"nova-metadata-0\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.174644 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.446064 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d6l9d"] Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.449274 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.465480 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d6l9d"] Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.473761 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-utilities\") pod \"community-operators-d6l9d\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.473869 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-catalog-content\") pod \"community-operators-d6l9d\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.473897 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhwt7\" (UniqueName: \"kubernetes.io/projected/8958815f-bb66-4dcc-874d-b9457757b108-kube-api-access-xhwt7\") pod \"community-operators-d6l9d\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.497986 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35b456e0-f0a8-40d3-a8fe-77b9880cecc4" path="/var/lib/kubelet/pods/35b456e0-f0a8-40d3-a8fe-77b9880cecc4/volumes" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.499419 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d68b9a30-55fe-4d32-8cba-9a12dcdc210c" path="/var/lib/kubelet/pods/d68b9a30-55fe-4d32-8cba-9a12dcdc210c/volumes" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.581482 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-utilities\") pod \"community-operators-d6l9d\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.581636 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-catalog-content\") pod \"community-operators-d6l9d\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.581676 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhwt7\" (UniqueName: \"kubernetes.io/projected/8958815f-bb66-4dcc-874d-b9457757b108-kube-api-access-xhwt7\") pod \"community-operators-d6l9d\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.583280 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-utilities\") pod \"community-operators-d6l9d\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.583557 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-catalog-content\") pod \"community-operators-d6l9d\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.614726 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhwt7\" (UniqueName: \"kubernetes.io/projected/8958815f-bb66-4dcc-874d-b9457757b108-kube-api-access-xhwt7\") pod \"community-operators-d6l9d\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.645435 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:09 crc kubenswrapper[4719]: W1215 12:41:09.650056 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333a9883_5cc9_48cd_81cd_e1d87e91006c.slice/crio-004cc539ae11356117efd2c50a7ffb8b5c96ad496136edd9f1f849dc72c9546c WatchSource:0}: Error finding container 004cc539ae11356117efd2c50a7ffb8b5c96ad496136edd9f1f849dc72c9546c: Status 404 returned error can't find the container with id 004cc539ae11356117efd2c50a7ffb8b5c96ad496136edd9f1f849dc72c9546c Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.747782 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"333a9883-5cc9-48cd-81cd-e1d87e91006c","Type":"ContainerStarted","Data":"004cc539ae11356117efd2c50a7ffb8b5c96ad496136edd9f1f849dc72c9546c"} Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.750557 4719 generic.go:334] "Generic (PLEG): container finished" podID="30929e15-1877-4485-9366-6396d32df2ae" containerID="166b49effe874c8c992db325f19173be0def7b459985ad660ca508d4805b4f8f" exitCode=0 Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.750598 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vzsnw" event={"ID":"30929e15-1877-4485-9366-6396d32df2ae","Type":"ContainerDied","Data":"166b49effe874c8c992db325f19173be0def7b459985ad660ca508d4805b4f8f"} Dec 15 12:41:09 crc kubenswrapper[4719]: I1215 12:41:09.775628 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:10 crc kubenswrapper[4719]: I1215 12:41:10.316020 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d6l9d"] Dec 15 12:41:10 crc kubenswrapper[4719]: W1215 12:41:10.316469 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8958815f_bb66_4dcc_874d_b9457757b108.slice/crio-653accd105d763a20cdb71fcdf3f4e37eed6eef27f35831c8d502c260b5fa69b WatchSource:0}: Error finding container 653accd105d763a20cdb71fcdf3f4e37eed6eef27f35831c8d502c260b5fa69b: Status 404 returned error can't find the container with id 653accd105d763a20cdb71fcdf3f4e37eed6eef27f35831c8d502c260b5fa69b Dec 15 12:41:10 crc kubenswrapper[4719]: I1215 12:41:10.763216 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"333a9883-5cc9-48cd-81cd-e1d87e91006c","Type":"ContainerStarted","Data":"5c4f0e096c22f396cf0d005430ac53706e6b6428ae063ea53ffeb327d82b6f2d"} Dec 15 12:41:10 crc kubenswrapper[4719]: I1215 12:41:10.763567 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"333a9883-5cc9-48cd-81cd-e1d87e91006c","Type":"ContainerStarted","Data":"344c544c0dfbe421682c5a330b75614cc51486e3e0379d87b56e53f0a4dfa119"} Dec 15 12:41:10 crc kubenswrapper[4719]: I1215 12:41:10.764785 4719 generic.go:334] "Generic (PLEG): container finished" podID="8958815f-bb66-4dcc-874d-b9457757b108" containerID="7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625" exitCode=0 Dec 15 12:41:10 crc kubenswrapper[4719]: I1215 12:41:10.764898 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d6l9d" event={"ID":"8958815f-bb66-4dcc-874d-b9457757b108","Type":"ContainerDied","Data":"7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625"} Dec 15 12:41:10 crc kubenswrapper[4719]: I1215 12:41:10.765013 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d6l9d" event={"ID":"8958815f-bb66-4dcc-874d-b9457757b108","Type":"ContainerStarted","Data":"653accd105d763a20cdb71fcdf3f4e37eed6eef27f35831c8d502c260b5fa69b"} Dec 15 12:41:10 crc kubenswrapper[4719]: I1215 12:41:10.827210 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.827191431 podStartE2EDuration="2.827191431s" podCreationTimestamp="2025-12-15 12:41:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:41:10.794154473 +0000 UTC m=+1431.736447503" watchObservedRunningTime="2025-12-15 12:41:10.827191431 +0000 UTC m=+1431.769484461" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.150757 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.313298 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fj4z6\" (UniqueName: \"kubernetes.io/projected/30929e15-1877-4485-9366-6396d32df2ae-kube-api-access-fj4z6\") pod \"30929e15-1877-4485-9366-6396d32df2ae\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.314447 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-combined-ca-bundle\") pod \"30929e15-1877-4485-9366-6396d32df2ae\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.314524 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-config-data\") pod \"30929e15-1877-4485-9366-6396d32df2ae\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.314548 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-scripts\") pod \"30929e15-1877-4485-9366-6396d32df2ae\" (UID: \"30929e15-1877-4485-9366-6396d32df2ae\") " Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.319171 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-scripts" (OuterVolumeSpecName: "scripts") pod "30929e15-1877-4485-9366-6396d32df2ae" (UID: "30929e15-1877-4485-9366-6396d32df2ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.321013 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30929e15-1877-4485-9366-6396d32df2ae-kube-api-access-fj4z6" (OuterVolumeSpecName: "kube-api-access-fj4z6") pod "30929e15-1877-4485-9366-6396d32df2ae" (UID: "30929e15-1877-4485-9366-6396d32df2ae"). InnerVolumeSpecName "kube-api-access-fj4z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.344066 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-config-data" (OuterVolumeSpecName: "config-data") pod "30929e15-1877-4485-9366-6396d32df2ae" (UID: "30929e15-1877-4485-9366-6396d32df2ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.353424 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30929e15-1877-4485-9366-6396d32df2ae" (UID: "30929e15-1877-4485-9366-6396d32df2ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.417611 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fj4z6\" (UniqueName: \"kubernetes.io/projected/30929e15-1877-4485-9366-6396d32df2ae-kube-api-access-fj4z6\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.417641 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.417650 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.417658 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30929e15-1877-4485-9366-6396d32df2ae-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.774259 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vzsnw" event={"ID":"30929e15-1877-4485-9366-6396d32df2ae","Type":"ContainerDied","Data":"e227a2c07eba898614ec6e8399d0b16edf2179f2e52f9dea187372a134123143"} Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.774302 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e227a2c07eba898614ec6e8399d0b16edf2179f2e52f9dea187372a134123143" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.774272 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vzsnw" Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.959155 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.959598 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerName="nova-api-log" containerID="cri-o://bd6c8d4b6435307bdf79d1759e4f8dfe321baf21ca202c9f61f0e5191456e363" gracePeriod=30 Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.959651 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerName="nova-api-api" containerID="cri-o://873d279897789b83274cf74d5e7acd31b8cebc22b0e832ae5fc624a833e1c844" gracePeriod=30 Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.975674 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:41:11 crc kubenswrapper[4719]: I1215 12:41:11.975928 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="40707f09-eaf4-4dfc-bd8c-e4d1336c327b" containerName="nova-scheduler-scheduler" containerID="cri-o://d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d" gracePeriod=30 Dec 15 12:41:12 crc kubenswrapper[4719]: I1215 12:41:12.031657 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:12 crc kubenswrapper[4719]: I1215 12:41:12.785607 4719 generic.go:334] "Generic (PLEG): container finished" podID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerID="bd6c8d4b6435307bdf79d1759e4f8dfe321baf21ca202c9f61f0e5191456e363" exitCode=143 Dec 15 12:41:12 crc kubenswrapper[4719]: I1215 12:41:12.785784 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5c62a0a4-152d-4953-a2c3-ab9343c208ba","Type":"ContainerDied","Data":"bd6c8d4b6435307bdf79d1759e4f8dfe321baf21ca202c9f61f0e5191456e363"} Dec 15 12:41:12 crc kubenswrapper[4719]: I1215 12:41:12.787616 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d6l9d" event={"ID":"8958815f-bb66-4dcc-874d-b9457757b108","Type":"ContainerStarted","Data":"2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725"} Dec 15 12:41:12 crc kubenswrapper[4719]: I1215 12:41:12.787706 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="333a9883-5cc9-48cd-81cd-e1d87e91006c" containerName="nova-metadata-log" containerID="cri-o://344c544c0dfbe421682c5a330b75614cc51486e3e0379d87b56e53f0a4dfa119" gracePeriod=30 Dec 15 12:41:12 crc kubenswrapper[4719]: I1215 12:41:12.788234 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="333a9883-5cc9-48cd-81cd-e1d87e91006c" containerName="nova-metadata-metadata" containerID="cri-o://5c4f0e096c22f396cf0d005430ac53706e6b6428ae063ea53ffeb327d82b6f2d" gracePeriod=30 Dec 15 12:41:13 crc kubenswrapper[4719]: I1215 12:41:13.031380 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5784cf869f-b4mfs" podUID="d68b9a30-55fe-4d32-8cba-9a12dcdc210c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.162:5353: i/o timeout" Dec 15 12:41:13 crc kubenswrapper[4719]: I1215 12:41:13.806535 4719 generic.go:334] "Generic (PLEG): container finished" podID="333a9883-5cc9-48cd-81cd-e1d87e91006c" containerID="5c4f0e096c22f396cf0d005430ac53706e6b6428ae063ea53ffeb327d82b6f2d" exitCode=0 Dec 15 12:41:13 crc kubenswrapper[4719]: I1215 12:41:13.806836 4719 generic.go:334] "Generic (PLEG): container finished" podID="333a9883-5cc9-48cd-81cd-e1d87e91006c" containerID="344c544c0dfbe421682c5a330b75614cc51486e3e0379d87b56e53f0a4dfa119" exitCode=143 Dec 15 12:41:13 crc kubenswrapper[4719]: I1215 12:41:13.806632 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"333a9883-5cc9-48cd-81cd-e1d87e91006c","Type":"ContainerDied","Data":"5c4f0e096c22f396cf0d005430ac53706e6b6428ae063ea53ffeb327d82b6f2d"} Dec 15 12:41:13 crc kubenswrapper[4719]: I1215 12:41:13.806926 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"333a9883-5cc9-48cd-81cd-e1d87e91006c","Type":"ContainerDied","Data":"344c544c0dfbe421682c5a330b75614cc51486e3e0379d87b56e53f0a4dfa119"} Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.175944 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.176189 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.297690 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.373270 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-nova-metadata-tls-certs\") pod \"333a9883-5cc9-48cd-81cd-e1d87e91006c\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.373517 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-combined-ca-bundle\") pod \"333a9883-5cc9-48cd-81cd-e1d87e91006c\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.373646 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-config-data\") pod \"333a9883-5cc9-48cd-81cd-e1d87e91006c\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.373795 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tkrx\" (UniqueName: \"kubernetes.io/projected/333a9883-5cc9-48cd-81cd-e1d87e91006c-kube-api-access-4tkrx\") pod \"333a9883-5cc9-48cd-81cd-e1d87e91006c\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.374041 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/333a9883-5cc9-48cd-81cd-e1d87e91006c-logs\") pod \"333a9883-5cc9-48cd-81cd-e1d87e91006c\" (UID: \"333a9883-5cc9-48cd-81cd-e1d87e91006c\") " Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.374399 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/333a9883-5cc9-48cd-81cd-e1d87e91006c-logs" (OuterVolumeSpecName: "logs") pod "333a9883-5cc9-48cd-81cd-e1d87e91006c" (UID: "333a9883-5cc9-48cd-81cd-e1d87e91006c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.374636 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/333a9883-5cc9-48cd-81cd-e1d87e91006c-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.385039 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/333a9883-5cc9-48cd-81cd-e1d87e91006c-kube-api-access-4tkrx" (OuterVolumeSpecName: "kube-api-access-4tkrx") pod "333a9883-5cc9-48cd-81cd-e1d87e91006c" (UID: "333a9883-5cc9-48cd-81cd-e1d87e91006c"). InnerVolumeSpecName "kube-api-access-4tkrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.416095 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-config-data" (OuterVolumeSpecName: "config-data") pod "333a9883-5cc9-48cd-81cd-e1d87e91006c" (UID: "333a9883-5cc9-48cd-81cd-e1d87e91006c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.419603 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "333a9883-5cc9-48cd-81cd-e1d87e91006c" (UID: "333a9883-5cc9-48cd-81cd-e1d87e91006c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.465135 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "333a9883-5cc9-48cd-81cd-e1d87e91006c" (UID: "333a9883-5cc9-48cd-81cd-e1d87e91006c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.476064 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tkrx\" (UniqueName: \"kubernetes.io/projected/333a9883-5cc9-48cd-81cd-e1d87e91006c-kube-api-access-4tkrx\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.476093 4719 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.476103 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.476111 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/333a9883-5cc9-48cd-81cd-e1d87e91006c-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.593996 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.781431 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-config-data\") pod \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.781522 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-combined-ca-bundle\") pod \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.781565 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79wlj\" (UniqueName: \"kubernetes.io/projected/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-kube-api-access-79wlj\") pod \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\" (UID: \"40707f09-eaf4-4dfc-bd8c-e4d1336c327b\") " Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.802833 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-kube-api-access-79wlj" (OuterVolumeSpecName: "kube-api-access-79wlj") pod "40707f09-eaf4-4dfc-bd8c-e4d1336c327b" (UID: "40707f09-eaf4-4dfc-bd8c-e4d1336c327b"). InnerVolumeSpecName "kube-api-access-79wlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.818926 4719 generic.go:334] "Generic (PLEG): container finished" podID="40707f09-eaf4-4dfc-bd8c-e4d1336c327b" containerID="d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d" exitCode=0 Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.819211 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.819116 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"40707f09-eaf4-4dfc-bd8c-e4d1336c327b","Type":"ContainerDied","Data":"d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d"} Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.819275 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"40707f09-eaf4-4dfc-bd8c-e4d1336c327b","Type":"ContainerDied","Data":"a7c0f25f95f8b8d9fbb100f1242cebd2b9ce58de7e8688fb50e8c7f37342c77f"} Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.819304 4719 scope.go:117] "RemoveContainer" containerID="d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.822132 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"333a9883-5cc9-48cd-81cd-e1d87e91006c","Type":"ContainerDied","Data":"004cc539ae11356117efd2c50a7ffb8b5c96ad496136edd9f1f849dc72c9546c"} Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.822231 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.826952 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40707f09-eaf4-4dfc-bd8c-e4d1336c327b" (UID: "40707f09-eaf4-4dfc-bd8c-e4d1336c327b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.828753 4719 generic.go:334] "Generic (PLEG): container finished" podID="8958815f-bb66-4dcc-874d-b9457757b108" containerID="2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725" exitCode=0 Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.828797 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d6l9d" event={"ID":"8958815f-bb66-4dcc-874d-b9457757b108","Type":"ContainerDied","Data":"2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725"} Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.860016 4719 scope.go:117] "RemoveContainer" containerID="d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d" Dec 15 12:41:14 crc kubenswrapper[4719]: E1215 12:41:14.863508 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d\": container with ID starting with d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d not found: ID does not exist" containerID="d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.863544 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d"} err="failed to get container status \"d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d\": rpc error: code = NotFound desc = could not find container \"d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d\": container with ID starting with d6c734511dc015d204d9e10bf23dca9f064a12d68f735b2b7efe1444bdf9fb6d not found: ID does not exist" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.863569 4719 scope.go:117] "RemoveContainer" containerID="5c4f0e096c22f396cf0d005430ac53706e6b6428ae063ea53ffeb327d82b6f2d" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.880150 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-config-data" (OuterVolumeSpecName: "config-data") pod "40707f09-eaf4-4dfc-bd8c-e4d1336c327b" (UID: "40707f09-eaf4-4dfc-bd8c-e4d1336c327b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.894390 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.894665 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.894971 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79wlj\" (UniqueName: \"kubernetes.io/projected/40707f09-eaf4-4dfc-bd8c-e4d1336c327b-kube-api-access-79wlj\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.911916 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.923981 4719 scope.go:117] "RemoveContainer" containerID="344c544c0dfbe421682c5a330b75614cc51486e3e0379d87b56e53f0a4dfa119" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.934207 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.981540 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:14 crc kubenswrapper[4719]: E1215 12:41:14.982012 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333a9883-5cc9-48cd-81cd-e1d87e91006c" containerName="nova-metadata-metadata" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.982030 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="333a9883-5cc9-48cd-81cd-e1d87e91006c" containerName="nova-metadata-metadata" Dec 15 12:41:14 crc kubenswrapper[4719]: E1215 12:41:14.982061 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333a9883-5cc9-48cd-81cd-e1d87e91006c" containerName="nova-metadata-log" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.982069 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="333a9883-5cc9-48cd-81cd-e1d87e91006c" containerName="nova-metadata-log" Dec 15 12:41:14 crc kubenswrapper[4719]: E1215 12:41:14.982085 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40707f09-eaf4-4dfc-bd8c-e4d1336c327b" containerName="nova-scheduler-scheduler" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.982093 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="40707f09-eaf4-4dfc-bd8c-e4d1336c327b" containerName="nova-scheduler-scheduler" Dec 15 12:41:14 crc kubenswrapper[4719]: E1215 12:41:14.982107 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30929e15-1877-4485-9366-6396d32df2ae" containerName="nova-manage" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.982116 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="30929e15-1877-4485-9366-6396d32df2ae" containerName="nova-manage" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.982340 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="40707f09-eaf4-4dfc-bd8c-e4d1336c327b" containerName="nova-scheduler-scheduler" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.982358 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="333a9883-5cc9-48cd-81cd-e1d87e91006c" containerName="nova-metadata-log" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.982381 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="30929e15-1877-4485-9366-6396d32df2ae" containerName="nova-manage" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.982390 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="333a9883-5cc9-48cd-81cd-e1d87e91006c" containerName="nova-metadata-metadata" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.988309 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.988411 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.992489 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 15 12:41:14 crc kubenswrapper[4719]: I1215 12:41:14.992703 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.101123 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zrn5\" (UniqueName: \"kubernetes.io/projected/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-kube-api-access-9zrn5\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.101168 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.101205 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-config-data\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.101366 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-logs\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.101550 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.157977 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.174091 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.186993 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.188323 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.213907 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.219889 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-config-data\") pod \"nova-scheduler-0\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.219952 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-logs\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.220034 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.220077 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.220188 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xmld\" (UniqueName: \"kubernetes.io/projected/be6799a7-1506-4f57-abba-a0db463847c9-kube-api-access-6xmld\") pod \"nova-scheduler-0\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.220232 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zrn5\" (UniqueName: \"kubernetes.io/projected/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-kube-api-access-9zrn5\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.220291 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.220335 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-config-data\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.223958 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-logs\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.229627 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-config-data\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.241533 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.247663 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.251829 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.287001 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zrn5\" (UniqueName: \"kubernetes.io/projected/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-kube-api-access-9zrn5\") pod \"nova-metadata-0\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.319674 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.325657 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.326028 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xmld\" (UniqueName: \"kubernetes.io/projected/be6799a7-1506-4f57-abba-a0db463847c9-kube-api-access-6xmld\") pod \"nova-scheduler-0\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.326174 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-config-data\") pod \"nova-scheduler-0\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.331665 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-config-data\") pod \"nova-scheduler-0\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.337658 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.360805 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xmld\" (UniqueName: \"kubernetes.io/projected/be6799a7-1506-4f57-abba-a0db463847c9-kube-api-access-6xmld\") pod \"nova-scheduler-0\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.486210 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.525784 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="333a9883-5cc9-48cd-81cd-e1d87e91006c" path="/var/lib/kubelet/pods/333a9883-5cc9-48cd-81cd-e1d87e91006c/volumes" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.526885 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40707f09-eaf4-4dfc-bd8c-e4d1336c327b" path="/var/lib/kubelet/pods/40707f09-eaf4-4dfc-bd8c-e4d1336c327b/volumes" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.879277 4719 generic.go:334] "Generic (PLEG): container finished" podID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerID="873d279897789b83274cf74d5e7acd31b8cebc22b0e832ae5fc624a833e1c844" exitCode=0 Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.879659 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5c62a0a4-152d-4953-a2c3-ab9343c208ba","Type":"ContainerDied","Data":"873d279897789b83274cf74d5e7acd31b8cebc22b0e832ae5fc624a833e1c844"} Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.891338 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d6l9d" event={"ID":"8958815f-bb66-4dcc-874d-b9457757b108","Type":"ContainerStarted","Data":"2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f"} Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.924098 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d6l9d" podStartSLOduration=2.201668688 podStartE2EDuration="6.924081485s" podCreationTimestamp="2025-12-15 12:41:09 +0000 UTC" firstStartedPulling="2025-12-15 12:41:10.766739752 +0000 UTC m=+1431.709032782" lastFinishedPulling="2025-12-15 12:41:15.489152559 +0000 UTC m=+1436.431445579" observedRunningTime="2025-12-15 12:41:15.923793626 +0000 UTC m=+1436.866086656" watchObservedRunningTime="2025-12-15 12:41:15.924081485 +0000 UTC m=+1436.866374515" Dec 15 12:41:15 crc kubenswrapper[4719]: I1215 12:41:15.960153 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.059646 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.141799 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-config-data\") pod \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.141890 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-combined-ca-bundle\") pod \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.141922 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c62a0a4-152d-4953-a2c3-ab9343c208ba-logs\") pod \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.142040 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhgsp\" (UniqueName: \"kubernetes.io/projected/5c62a0a4-152d-4953-a2c3-ab9343c208ba-kube-api-access-bhgsp\") pod \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\" (UID: \"5c62a0a4-152d-4953-a2c3-ab9343c208ba\") " Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.143210 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c62a0a4-152d-4953-a2c3-ab9343c208ba-logs" (OuterVolumeSpecName: "logs") pod "5c62a0a4-152d-4953-a2c3-ab9343c208ba" (UID: "5c62a0a4-152d-4953-a2c3-ab9343c208ba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.149361 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c62a0a4-152d-4953-a2c3-ab9343c208ba-kube-api-access-bhgsp" (OuterVolumeSpecName: "kube-api-access-bhgsp") pod "5c62a0a4-152d-4953-a2c3-ab9343c208ba" (UID: "5c62a0a4-152d-4953-a2c3-ab9343c208ba"). InnerVolumeSpecName "kube-api-access-bhgsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.174030 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-config-data" (OuterVolumeSpecName: "config-data") pod "5c62a0a4-152d-4953-a2c3-ab9343c208ba" (UID: "5c62a0a4-152d-4953-a2c3-ab9343c208ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.181736 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c62a0a4-152d-4953-a2c3-ab9343c208ba" (UID: "5c62a0a4-152d-4953-a2c3-ab9343c208ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.215536 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.244484 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.244517 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c62a0a4-152d-4953-a2c3-ab9343c208ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.244531 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c62a0a4-152d-4953-a2c3-ab9343c208ba-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.244542 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhgsp\" (UniqueName: \"kubernetes.io/projected/5c62a0a4-152d-4953-a2c3-ab9343c208ba-kube-api-access-bhgsp\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.901307 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be6799a7-1506-4f57-abba-a0db463847c9","Type":"ContainerStarted","Data":"c93b14720c53b973f522fbca48014e3e3894a38553fcaa0d853e0d6adbb90741"} Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.903458 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80","Type":"ContainerStarted","Data":"c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637"} Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.903489 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80","Type":"ContainerStarted","Data":"e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be"} Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.903503 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80","Type":"ContainerStarted","Data":"6c5d65692b93dd1a96b94a112de2a4349eabbeccd94613abfe2278e76ccb0c3d"} Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.905268 4719 generic.go:334] "Generic (PLEG): container finished" podID="1c9e9ff2-3aa8-4640-979a-be62e16eca47" containerID="4504d72e3231d8806da3f3047b136c2e704fefb6a67479e700f38fc11bc789df" exitCode=0 Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.905346 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5p65f" event={"ID":"1c9e9ff2-3aa8-4640-979a-be62e16eca47","Type":"ContainerDied","Data":"4504d72e3231d8806da3f3047b136c2e704fefb6a67479e700f38fc11bc789df"} Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.907303 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5c62a0a4-152d-4953-a2c3-ab9343c208ba","Type":"ContainerDied","Data":"2e903192991901c1042ad1e32b98ad36af2453cf54f87b6fbbabc14b48c71e6c"} Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.907334 4719 scope.go:117] "RemoveContainer" containerID="873d279897789b83274cf74d5e7acd31b8cebc22b0e832ae5fc624a833e1c844" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.907338 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.954761 4719 scope.go:117] "RemoveContainer" containerID="bd6c8d4b6435307bdf79d1759e4f8dfe321baf21ca202c9f61f0e5191456e363" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.960182 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.981805 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.997984 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:16 crc kubenswrapper[4719]: E1215 12:41:16.998373 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerName="nova-api-log" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.998385 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerName="nova-api-log" Dec 15 12:41:16 crc kubenswrapper[4719]: E1215 12:41:16.998406 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerName="nova-api-api" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.998412 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerName="nova-api-api" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.998604 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerName="nova-api-api" Dec 15 12:41:16 crc kubenswrapper[4719]: I1215 12:41:16.998622 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" containerName="nova-api-log" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:16.999605 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.005802 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.006785 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.163047 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.163114 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572d65f8-13e4-4e29-b7e1-84d489b0acb0-logs\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.163132 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzf7p\" (UniqueName: \"kubernetes.io/projected/572d65f8-13e4-4e29-b7e1-84d489b0acb0-kube-api-access-mzf7p\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.163180 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-config-data\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.264723 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-config-data\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.264872 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.264909 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572d65f8-13e4-4e29-b7e1-84d489b0acb0-logs\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.264939 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzf7p\" (UniqueName: \"kubernetes.io/projected/572d65f8-13e4-4e29-b7e1-84d489b0acb0-kube-api-access-mzf7p\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.266160 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572d65f8-13e4-4e29-b7e1-84d489b0acb0-logs\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.272392 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-config-data\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.274396 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.285616 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzf7p\" (UniqueName: \"kubernetes.io/projected/572d65f8-13e4-4e29-b7e1-84d489b0acb0-kube-api-access-mzf7p\") pod \"nova-api-0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.320070 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.519739 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c62a0a4-152d-4953-a2c3-ab9343c208ba" path="/var/lib/kubelet/pods/5c62a0a4-152d-4953-a2c3-ab9343c208ba/volumes" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.843185 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.920289 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"572d65f8-13e4-4e29-b7e1-84d489b0acb0","Type":"ContainerStarted","Data":"59f2423eda0c2a6ad7a4e71909ec2d5b619ce4d6d3f5a706f081f85a68b71d73"} Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.926101 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be6799a7-1506-4f57-abba-a0db463847c9","Type":"ContainerStarted","Data":"17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625"} Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.963739 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.963718985 podStartE2EDuration="2.963718985s" podCreationTimestamp="2025-12-15 12:41:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:41:17.957060946 +0000 UTC m=+1438.899353976" watchObservedRunningTime="2025-12-15 12:41:17.963718985 +0000 UTC m=+1438.906012015" Dec 15 12:41:17 crc kubenswrapper[4719]: I1215 12:41:17.982598 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.982579348 podStartE2EDuration="3.982579348s" podCreationTimestamp="2025-12-15 12:41:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:41:17.97978178 +0000 UTC m=+1438.922074810" watchObservedRunningTime="2025-12-15 12:41:17.982579348 +0000 UTC m=+1438.924872378" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.113627 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vsgzc" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="registry-server" probeResult="failure" output=< Dec 15 12:41:18 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 12:41:18 crc kubenswrapper[4719]: > Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.399403 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.495067 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-scripts\") pod \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.495320 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-combined-ca-bundle\") pod \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.495369 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-config-data\") pod \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.495407 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4z24\" (UniqueName: \"kubernetes.io/projected/1c9e9ff2-3aa8-4640-979a-be62e16eca47-kube-api-access-b4z24\") pod \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\" (UID: \"1c9e9ff2-3aa8-4640-979a-be62e16eca47\") " Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.500452 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-scripts" (OuterVolumeSpecName: "scripts") pod "1c9e9ff2-3aa8-4640-979a-be62e16eca47" (UID: "1c9e9ff2-3aa8-4640-979a-be62e16eca47"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.502661 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c9e9ff2-3aa8-4640-979a-be62e16eca47-kube-api-access-b4z24" (OuterVolumeSpecName: "kube-api-access-b4z24") pod "1c9e9ff2-3aa8-4640-979a-be62e16eca47" (UID: "1c9e9ff2-3aa8-4640-979a-be62e16eca47"). InnerVolumeSpecName "kube-api-access-b4z24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.523298 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-config-data" (OuterVolumeSpecName: "config-data") pod "1c9e9ff2-3aa8-4640-979a-be62e16eca47" (UID: "1c9e9ff2-3aa8-4640-979a-be62e16eca47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.526088 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c9e9ff2-3aa8-4640-979a-be62e16eca47" (UID: "1c9e9ff2-3aa8-4640-979a-be62e16eca47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.597273 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.597554 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.597565 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4z24\" (UniqueName: \"kubernetes.io/projected/1c9e9ff2-3aa8-4640-979a-be62e16eca47-kube-api-access-b4z24\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.597574 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9e9ff2-3aa8-4640-979a-be62e16eca47-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.942443 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5p65f" event={"ID":"1c9e9ff2-3aa8-4640-979a-be62e16eca47","Type":"ContainerDied","Data":"f6eac6e18dcf460771584cd39cf1d0c794cc03adc1606e10a6868dc61531b517"} Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.942484 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6eac6e18dcf460771584cd39cf1d0c794cc03adc1606e10a6868dc61531b517" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.942554 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5p65f" Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.948161 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"572d65f8-13e4-4e29-b7e1-84d489b0acb0","Type":"ContainerStarted","Data":"b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b"} Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.948230 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"572d65f8-13e4-4e29-b7e1-84d489b0acb0","Type":"ContainerStarted","Data":"782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528"} Dec 15 12:41:18 crc kubenswrapper[4719]: I1215 12:41:18.994319 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.994289558 podStartE2EDuration="2.994289558s" podCreationTimestamp="2025-12-15 12:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:41:18.97841733 +0000 UTC m=+1439.920710360" watchObservedRunningTime="2025-12-15 12:41:18.994289558 +0000 UTC m=+1439.936582588" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.033888 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 15 12:41:19 crc kubenswrapper[4719]: E1215 12:41:19.034287 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9e9ff2-3aa8-4640-979a-be62e16eca47" containerName="nova-cell1-conductor-db-sync" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.034299 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9e9ff2-3aa8-4640-979a-be62e16eca47" containerName="nova-cell1-conductor-db-sync" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.034458 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9e9ff2-3aa8-4640-979a-be62e16eca47" containerName="nova-cell1-conductor-db-sync" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.048545 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.048666 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.053976 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.209448 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c7f0b4-980a-4543-a7bc-38e6115a9d53-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"36c7f0b4-980a-4543-a7bc-38e6115a9d53\") " pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.209569 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c7f0b4-980a-4543-a7bc-38e6115a9d53-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"36c7f0b4-980a-4543-a7bc-38e6115a9d53\") " pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.209730 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64gn4\" (UniqueName: \"kubernetes.io/projected/36c7f0b4-980a-4543-a7bc-38e6115a9d53-kube-api-access-64gn4\") pod \"nova-cell1-conductor-0\" (UID: \"36c7f0b4-980a-4543-a7bc-38e6115a9d53\") " pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.311281 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64gn4\" (UniqueName: \"kubernetes.io/projected/36c7f0b4-980a-4543-a7bc-38e6115a9d53-kube-api-access-64gn4\") pod \"nova-cell1-conductor-0\" (UID: \"36c7f0b4-980a-4543-a7bc-38e6115a9d53\") " pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.311354 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c7f0b4-980a-4543-a7bc-38e6115a9d53-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"36c7f0b4-980a-4543-a7bc-38e6115a9d53\") " pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.311435 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c7f0b4-980a-4543-a7bc-38e6115a9d53-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"36c7f0b4-980a-4543-a7bc-38e6115a9d53\") " pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.317523 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c7f0b4-980a-4543-a7bc-38e6115a9d53-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"36c7f0b4-980a-4543-a7bc-38e6115a9d53\") " pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.317745 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c7f0b4-980a-4543-a7bc-38e6115a9d53-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"36c7f0b4-980a-4543-a7bc-38e6115a9d53\") " pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.331549 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64gn4\" (UniqueName: \"kubernetes.io/projected/36c7f0b4-980a-4543-a7bc-38e6115a9d53-kube-api-access-64gn4\") pod \"nova-cell1-conductor-0\" (UID: \"36c7f0b4-980a-4543-a7bc-38e6115a9d53\") " pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.379127 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.776789 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.777443 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.916075 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 15 12:41:19 crc kubenswrapper[4719]: I1215 12:41:19.965682 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"36c7f0b4-980a-4543-a7bc-38e6115a9d53","Type":"ContainerStarted","Data":"7d2d846d50ed0ff4bd27f53a53c6ff05e18f7d6967b0dbcef08d1dab3bd2843e"} Dec 15 12:41:20 crc kubenswrapper[4719]: I1215 12:41:20.321160 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 15 12:41:20 crc kubenswrapper[4719]: I1215 12:41:20.321500 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 15 12:41:20 crc kubenswrapper[4719]: I1215 12:41:20.487745 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 15 12:41:20 crc kubenswrapper[4719]: I1215 12:41:20.842166 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d6l9d" podUID="8958815f-bb66-4dcc-874d-b9457757b108" containerName="registry-server" probeResult="failure" output=< Dec 15 12:41:20 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 12:41:20 crc kubenswrapper[4719]: > Dec 15 12:41:20 crc kubenswrapper[4719]: I1215 12:41:20.973298 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"36c7f0b4-980a-4543-a7bc-38e6115a9d53","Type":"ContainerStarted","Data":"3685f8a673d64572c3b933e3bee660f06fef106c2e35fb7c03353e13bc2edf83"} Dec 15 12:41:20 crc kubenswrapper[4719]: I1215 12:41:20.973466 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:20 crc kubenswrapper[4719]: I1215 12:41:20.992733 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.9927152430000001 podStartE2EDuration="1.992715243s" podCreationTimestamp="2025-12-15 12:41:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:41:20.989947186 +0000 UTC m=+1441.932240226" watchObservedRunningTime="2025-12-15 12:41:20.992715243 +0000 UTC m=+1441.935008273" Dec 15 12:41:24 crc kubenswrapper[4719]: I1215 12:41:24.141068 4719 patch_prober.go:28] interesting pod/console-5cf7dbcc69-vhz75 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 15 12:41:24 crc kubenswrapper[4719]: I1215 12:41:24.141670 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-5cf7dbcc69-vhz75" podUID="0340c677-46c9-46db-bc44-20372c6a4248" containerName="console" probeResult="failure" output="Get \"https://10.217.0.53:8443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:41:25 crc kubenswrapper[4719]: I1215 12:41:25.320936 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 15 12:41:25 crc kubenswrapper[4719]: I1215 12:41:25.321263 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 15 12:41:25 crc kubenswrapper[4719]: I1215 12:41:25.515828 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 15 12:41:25 crc kubenswrapper[4719]: I1215 12:41:25.635057 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 15 12:41:26 crc kubenswrapper[4719]: I1215 12:41:26.056457 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 15 12:41:26 crc kubenswrapper[4719]: I1215 12:41:26.333023 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:41:26 crc kubenswrapper[4719]: I1215 12:41:26.333023 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:41:27 crc kubenswrapper[4719]: I1215 12:41:27.088630 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:41:27 crc kubenswrapper[4719]: I1215 12:41:27.143370 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:41:27 crc kubenswrapper[4719]: I1215 12:41:27.320073 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 15 12:41:27 crc kubenswrapper[4719]: I1215 12:41:27.320113 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 15 12:41:27 crc kubenswrapper[4719]: I1215 12:41:27.326084 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vsgzc"] Dec 15 12:41:28 crc kubenswrapper[4719]: I1215 12:41:28.403119 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 15 12:41:28 crc kubenswrapper[4719]: I1215 12:41:28.403269 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 15 12:41:29 crc kubenswrapper[4719]: I1215 12:41:29.057582 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vsgzc" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="registry-server" containerID="cri-o://7de42768b6b795add51e739ccd2eec25ac3ad6203ca2dd4fbf459c81a96feea9" gracePeriod=2 Dec 15 12:41:29 crc kubenswrapper[4719]: I1215 12:41:29.415566 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 15 12:41:29 crc kubenswrapper[4719]: I1215 12:41:29.824894 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:29 crc kubenswrapper[4719]: I1215 12:41:29.872210 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.085692 4719 generic.go:334] "Generic (PLEG): container finished" podID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerID="7de42768b6b795add51e739ccd2eec25ac3ad6203ca2dd4fbf459c81a96feea9" exitCode=0 Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.085894 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsgzc" event={"ID":"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658","Type":"ContainerDied","Data":"7de42768b6b795add51e739ccd2eec25ac3ad6203ca2dd4fbf459c81a96feea9"} Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.258742 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.406119 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-utilities\") pod \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.406304 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7nml\" (UniqueName: \"kubernetes.io/projected/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-kube-api-access-r7nml\") pod \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.406376 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-catalog-content\") pod \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\" (UID: \"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658\") " Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.406992 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-utilities" (OuterVolumeSpecName: "utilities") pod "5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" (UID: "5ecc97bf-b6b8-4a59-a449-ca8b19c6f658"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.412775 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-kube-api-access-r7nml" (OuterVolumeSpecName: "kube-api-access-r7nml") pod "5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" (UID: "5ecc97bf-b6b8-4a59-a449-ca8b19c6f658"). InnerVolumeSpecName "kube-api-access-r7nml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.508133 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.508160 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7nml\" (UniqueName: \"kubernetes.io/projected/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-kube-api-access-r7nml\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.529196 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" (UID: "5ecc97bf-b6b8-4a59-a449-ca8b19c6f658"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:30 crc kubenswrapper[4719]: I1215 12:41:30.609571 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.101475 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsgzc" event={"ID":"5ecc97bf-b6b8-4a59-a449-ca8b19c6f658","Type":"ContainerDied","Data":"2dc9f622c00d9895fab4f748d96d7374598655704457cbc4d3756448b646a3ec"} Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.101533 4719 scope.go:117] "RemoveContainer" containerID="7de42768b6b795add51e739ccd2eec25ac3ad6203ca2dd4fbf459c81a96feea9" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.101683 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsgzc" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.140627 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d6l9d"] Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.140871 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d6l9d" podUID="8958815f-bb66-4dcc-874d-b9457757b108" containerName="registry-server" containerID="cri-o://2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f" gracePeriod=2 Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.152292 4719 scope.go:117] "RemoveContainer" containerID="32671fc64276bddfb76780a433ff0be436c090048a677a3f07ccd3cdc3f7ab23" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.168293 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vsgzc"] Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.186059 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vsgzc"] Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.188144 4719 scope.go:117] "RemoveContainer" containerID="875449d5418281b692ac2c63f1df67e3142449e6f43d331bc5f7112a5f9fa9c2" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.486152 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" path="/var/lib/kubelet/pods/5ecc97bf-b6b8-4a59-a449-ca8b19c6f658/volumes" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.623838 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.733496 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-utilities\") pod \"8958815f-bb66-4dcc-874d-b9457757b108\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.734680 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhwt7\" (UniqueName: \"kubernetes.io/projected/8958815f-bb66-4dcc-874d-b9457757b108-kube-api-access-xhwt7\") pod \"8958815f-bb66-4dcc-874d-b9457757b108\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.734623 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-utilities" (OuterVolumeSpecName: "utilities") pod "8958815f-bb66-4dcc-874d-b9457757b108" (UID: "8958815f-bb66-4dcc-874d-b9457757b108"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.735583 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-catalog-content\") pod \"8958815f-bb66-4dcc-874d-b9457757b108\" (UID: \"8958815f-bb66-4dcc-874d-b9457757b108\") " Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.736257 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.739207 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8958815f-bb66-4dcc-874d-b9457757b108-kube-api-access-xhwt7" (OuterVolumeSpecName: "kube-api-access-xhwt7") pod "8958815f-bb66-4dcc-874d-b9457757b108" (UID: "8958815f-bb66-4dcc-874d-b9457757b108"). InnerVolumeSpecName "kube-api-access-xhwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.781362 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8958815f-bb66-4dcc-874d-b9457757b108" (UID: "8958815f-bb66-4dcc-874d-b9457757b108"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.837665 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8958815f-bb66-4dcc-874d-b9457757b108-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:31 crc kubenswrapper[4719]: I1215 12:41:31.837700 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhwt7\" (UniqueName: \"kubernetes.io/projected/8958815f-bb66-4dcc-874d-b9457757b108-kube-api-access-xhwt7\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.111549 4719 generic.go:334] "Generic (PLEG): container finished" podID="8958815f-bb66-4dcc-874d-b9457757b108" containerID="2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f" exitCode=0 Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.111585 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d6l9d" event={"ID":"8958815f-bb66-4dcc-874d-b9457757b108","Type":"ContainerDied","Data":"2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f"} Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.111623 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d6l9d" event={"ID":"8958815f-bb66-4dcc-874d-b9457757b108","Type":"ContainerDied","Data":"653accd105d763a20cdb71fcdf3f4e37eed6eef27f35831c8d502c260b5fa69b"} Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.111677 4719 scope.go:117] "RemoveContainer" containerID="2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f" Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.111702 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d6l9d" Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.130613 4719 scope.go:117] "RemoveContainer" containerID="2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725" Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.144003 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d6l9d"] Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.151120 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d6l9d"] Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.168952 4719 scope.go:117] "RemoveContainer" containerID="7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625" Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.202074 4719 scope.go:117] "RemoveContainer" containerID="2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f" Dec 15 12:41:32 crc kubenswrapper[4719]: E1215 12:41:32.203100 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f\": container with ID starting with 2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f not found: ID does not exist" containerID="2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f" Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.203135 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f"} err="failed to get container status \"2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f\": rpc error: code = NotFound desc = could not find container \"2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f\": container with ID starting with 2603f4ad0bdefd65331ae1c8ceb05bde3a6cafcea48d84f99f6139f2b298ff5f not found: ID does not exist" Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.203155 4719 scope.go:117] "RemoveContainer" containerID="2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725" Dec 15 12:41:32 crc kubenswrapper[4719]: E1215 12:41:32.203483 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725\": container with ID starting with 2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725 not found: ID does not exist" containerID="2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725" Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.203579 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725"} err="failed to get container status \"2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725\": rpc error: code = NotFound desc = could not find container \"2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725\": container with ID starting with 2c47fa86ec6fd90f5e9a7d1fa57b724fd267475573055bd9ce80dcbfb6d63725 not found: ID does not exist" Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.203662 4719 scope.go:117] "RemoveContainer" containerID="7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625" Dec 15 12:41:32 crc kubenswrapper[4719]: E1215 12:41:32.204174 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625\": container with ID starting with 7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625 not found: ID does not exist" containerID="7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625" Dec 15 12:41:32 crc kubenswrapper[4719]: I1215 12:41:32.204199 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625"} err="failed to get container status \"7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625\": rpc error: code = NotFound desc = could not find container \"7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625\": container with ID starting with 7fc4e06b475b1024bbfef27a859eac83ee7ac95321c39f8620eb92161315b625 not found: ID does not exist" Dec 15 12:41:33 crc kubenswrapper[4719]: I1215 12:41:33.498795 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8958815f-bb66-4dcc-874d-b9457757b108" path="/var/lib/kubelet/pods/8958815f-bb66-4dcc-874d-b9457757b108/volumes" Dec 15 12:41:35 crc kubenswrapper[4719]: I1215 12:41:35.329969 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 15 12:41:35 crc kubenswrapper[4719]: I1215 12:41:35.332229 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 15 12:41:35 crc kubenswrapper[4719]: I1215 12:41:35.337670 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.071026 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.145792 4719 generic.go:334] "Generic (PLEG): container finished" podID="8ee6b204-39a3-4189-8ca8-8210ed849824" containerID="e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e" exitCode=137 Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.145947 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.146004 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8ee6b204-39a3-4189-8ca8-8210ed849824","Type":"ContainerDied","Data":"e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e"} Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.146044 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8ee6b204-39a3-4189-8ca8-8210ed849824","Type":"ContainerDied","Data":"0ee4c5fd432ab8a452541cb75785d19b7698682e78927cee127533842734bffd"} Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.146065 4719 scope.go:117] "RemoveContainer" containerID="e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.152706 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.170996 4719 scope.go:117] "RemoveContainer" containerID="e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e" Dec 15 12:41:36 crc kubenswrapper[4719]: E1215 12:41:36.171429 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e\": container with ID starting with e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e not found: ID does not exist" containerID="e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.171482 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e"} err="failed to get container status \"e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e\": rpc error: code = NotFound desc = could not find container \"e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e\": container with ID starting with e008599c5970a99cfbf8561bd805a5dbe0f693ddff2b25bb1b03deed8f99786e not found: ID does not exist" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.216696 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xg48\" (UniqueName: \"kubernetes.io/projected/8ee6b204-39a3-4189-8ca8-8210ed849824-kube-api-access-2xg48\") pod \"8ee6b204-39a3-4189-8ca8-8210ed849824\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.216825 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-combined-ca-bundle\") pod \"8ee6b204-39a3-4189-8ca8-8210ed849824\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.216898 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-config-data\") pod \"8ee6b204-39a3-4189-8ca8-8210ed849824\" (UID: \"8ee6b204-39a3-4189-8ca8-8210ed849824\") " Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.225497 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ee6b204-39a3-4189-8ca8-8210ed849824-kube-api-access-2xg48" (OuterVolumeSpecName: "kube-api-access-2xg48") pod "8ee6b204-39a3-4189-8ca8-8210ed849824" (UID: "8ee6b204-39a3-4189-8ca8-8210ed849824"). InnerVolumeSpecName "kube-api-access-2xg48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.258597 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-config-data" (OuterVolumeSpecName: "config-data") pod "8ee6b204-39a3-4189-8ca8-8210ed849824" (UID: "8ee6b204-39a3-4189-8ca8-8210ed849824"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.263986 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ee6b204-39a3-4189-8ca8-8210ed849824" (UID: "8ee6b204-39a3-4189-8ca8-8210ed849824"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.320536 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xg48\" (UniqueName: \"kubernetes.io/projected/8ee6b204-39a3-4189-8ca8-8210ed849824-kube-api-access-2xg48\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.320593 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.320603 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee6b204-39a3-4189-8ca8-8210ed849824-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.479347 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.490099 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.501180 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 15 12:41:36 crc kubenswrapper[4719]: E1215 12:41:36.501835 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8958815f-bb66-4dcc-874d-b9457757b108" containerName="registry-server" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.501880 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="8958815f-bb66-4dcc-874d-b9457757b108" containerName="registry-server" Dec 15 12:41:36 crc kubenswrapper[4719]: E1215 12:41:36.501899 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="extract-content" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.501908 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="extract-content" Dec 15 12:41:36 crc kubenswrapper[4719]: E1215 12:41:36.501934 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8958815f-bb66-4dcc-874d-b9457757b108" containerName="extract-utilities" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.501943 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="8958815f-bb66-4dcc-874d-b9457757b108" containerName="extract-utilities" Dec 15 12:41:36 crc kubenswrapper[4719]: E1215 12:41:36.501961 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ee6b204-39a3-4189-8ca8-8210ed849824" containerName="nova-cell1-novncproxy-novncproxy" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.501968 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ee6b204-39a3-4189-8ca8-8210ed849824" containerName="nova-cell1-novncproxy-novncproxy" Dec 15 12:41:36 crc kubenswrapper[4719]: E1215 12:41:36.501986 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="registry-server" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.501993 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="registry-server" Dec 15 12:41:36 crc kubenswrapper[4719]: E1215 12:41:36.502006 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="extract-utilities" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.502013 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="extract-utilities" Dec 15 12:41:36 crc kubenswrapper[4719]: E1215 12:41:36.502030 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8958815f-bb66-4dcc-874d-b9457757b108" containerName="extract-content" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.502036 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="8958815f-bb66-4dcc-874d-b9457757b108" containerName="extract-content" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.502227 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ee6b204-39a3-4189-8ca8-8210ed849824" containerName="nova-cell1-novncproxy-novncproxy" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.502261 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ecc97bf-b6b8-4a59-a449-ca8b19c6f658" containerName="registry-server" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.502279 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="8958815f-bb66-4dcc-874d-b9457757b108" containerName="registry-server" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.503064 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.507622 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.507962 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.507981 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.522455 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.523197 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.523284 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8vrl\" (UniqueName: \"kubernetes.io/projected/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-kube-api-access-v8vrl\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.523362 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.523489 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.523534 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.624599 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.624651 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.624718 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.624759 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8vrl\" (UniqueName: \"kubernetes.io/projected/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-kube-api-access-v8vrl\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.624811 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.629419 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.629440 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.629931 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.630591 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.646525 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8vrl\" (UniqueName: \"kubernetes.io/projected/ccc69b5f-a98b-4188-b08f-4cb6f58c085c-kube-api-access-v8vrl\") pod \"nova-cell1-novncproxy-0\" (UID: \"ccc69b5f-a98b-4188-b08f-4cb6f58c085c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:36 crc kubenswrapper[4719]: I1215 12:41:36.818608 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:37 crc kubenswrapper[4719]: I1215 12:41:37.318340 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 15 12:41:37 crc kubenswrapper[4719]: I1215 12:41:37.327830 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 15 12:41:37 crc kubenswrapper[4719]: I1215 12:41:37.329364 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 15 12:41:37 crc kubenswrapper[4719]: I1215 12:41:37.329465 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 15 12:41:37 crc kubenswrapper[4719]: I1215 12:41:37.333016 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 15 12:41:37 crc kubenswrapper[4719]: I1215 12:41:37.486612 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ee6b204-39a3-4189-8ca8-8210ed849824" path="/var/lib/kubelet/pods/8ee6b204-39a3-4189-8ca8-8210ed849824/volumes" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.172730 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ccc69b5f-a98b-4188-b08f-4cb6f58c085c","Type":"ContainerStarted","Data":"9087a5267bda303c5b562fdd1af8ad9ad9e11f52c338a19243e503b9077bbb99"} Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.173492 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.173516 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ccc69b5f-a98b-4188-b08f-4cb6f58c085c","Type":"ContainerStarted","Data":"216af9bfd44bdd4c10c790ffda957ea13afa86d79c3f323d5544fe89351072bc"} Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.178129 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.194219 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.194205102 podStartE2EDuration="2.194205102s" podCreationTimestamp="2025-12-15 12:41:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:41:38.193154228 +0000 UTC m=+1459.135447258" watchObservedRunningTime="2025-12-15 12:41:38.194205102 +0000 UTC m=+1459.136498122" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.396513 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-psnpv"] Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.398152 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.425524 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-psnpv"] Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.598595 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.598651 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.599048 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.599234 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.599337 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdcws\" (UniqueName: \"kubernetes.io/projected/2fd5121b-dd23-4935-a6d2-3a81e92193cf-kube-api-access-jdcws\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.599491 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-config\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.701330 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.701652 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.701692 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdcws\" (UniqueName: \"kubernetes.io/projected/2fd5121b-dd23-4935-a6d2-3a81e92193cf-kube-api-access-jdcws\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.701745 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-config\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.701829 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.701872 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.702683 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.703229 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.703715 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.704506 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-config\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.705050 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:38 crc kubenswrapper[4719]: I1215 12:41:38.734826 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdcws\" (UniqueName: \"kubernetes.io/projected/2fd5121b-dd23-4935-a6d2-3a81e92193cf-kube-api-access-jdcws\") pod \"dnsmasq-dns-59cf4bdb65-psnpv\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:39 crc kubenswrapper[4719]: I1215 12:41:39.016556 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:39 crc kubenswrapper[4719]: I1215 12:41:39.596644 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-psnpv"] Dec 15 12:41:40 crc kubenswrapper[4719]: I1215 12:41:40.188832 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" event={"ID":"2fd5121b-dd23-4935-a6d2-3a81e92193cf","Type":"ContainerStarted","Data":"92088af17ad5bcfff853a8824caab0b37d1a3265a349883a837d4a7c262a0050"} Dec 15 12:41:41 crc kubenswrapper[4719]: I1215 12:41:41.118398 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:41 crc kubenswrapper[4719]: I1215 12:41:41.195476 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerName="nova-api-log" containerID="cri-o://782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528" gracePeriod=30 Dec 15 12:41:41 crc kubenswrapper[4719]: I1215 12:41:41.195557 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerName="nova-api-api" containerID="cri-o://b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b" gracePeriod=30 Dec 15 12:41:41 crc kubenswrapper[4719]: I1215 12:41:41.819650 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:42 crc kubenswrapper[4719]: I1215 12:41:42.204962 4719 generic.go:334] "Generic (PLEG): container finished" podID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerID="782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528" exitCode=143 Dec 15 12:41:42 crc kubenswrapper[4719]: I1215 12:41:42.205026 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"572d65f8-13e4-4e29-b7e1-84d489b0acb0","Type":"ContainerDied","Data":"782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528"} Dec 15 12:41:42 crc kubenswrapper[4719]: I1215 12:41:42.207024 4719 generic.go:334] "Generic (PLEG): container finished" podID="2fd5121b-dd23-4935-a6d2-3a81e92193cf" containerID="077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a" exitCode=0 Dec 15 12:41:42 crc kubenswrapper[4719]: I1215 12:41:42.207051 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" event={"ID":"2fd5121b-dd23-4935-a6d2-3a81e92193cf","Type":"ContainerDied","Data":"077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a"} Dec 15 12:41:42 crc kubenswrapper[4719]: I1215 12:41:42.470753 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:41:42 crc kubenswrapper[4719]: I1215 12:41:42.471398 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="ceilometer-central-agent" containerID="cri-o://0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75" gracePeriod=30 Dec 15 12:41:42 crc kubenswrapper[4719]: I1215 12:41:42.471530 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="ceilometer-notification-agent" containerID="cri-o://d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54" gracePeriod=30 Dec 15 12:41:42 crc kubenswrapper[4719]: I1215 12:41:42.471606 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="proxy-httpd" containerID="cri-o://1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f" gracePeriod=30 Dec 15 12:41:42 crc kubenswrapper[4719]: I1215 12:41:42.471634 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="sg-core" containerID="cri-o://2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8" gracePeriod=30 Dec 15 12:41:43 crc kubenswrapper[4719]: E1215 12:41:43.025298 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde640a25_d092_4c73_bb2f_c7a21993261e.slice/crio-0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde640a25_d092_4c73_bb2f_c7a21993261e.slice/crio-conmon-0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75.scope\": RecentStats: unable to find data in memory cache]" Dec 15 12:41:43 crc kubenswrapper[4719]: I1215 12:41:43.220561 4719 generic.go:334] "Generic (PLEG): container finished" podID="de640a25-d092-4c73-bb2f-c7a21993261e" containerID="1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f" exitCode=0 Dec 15 12:41:43 crc kubenswrapper[4719]: I1215 12:41:43.220594 4719 generic.go:334] "Generic (PLEG): container finished" podID="de640a25-d092-4c73-bb2f-c7a21993261e" containerID="2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8" exitCode=2 Dec 15 12:41:43 crc kubenswrapper[4719]: I1215 12:41:43.220603 4719 generic.go:334] "Generic (PLEG): container finished" podID="de640a25-d092-4c73-bb2f-c7a21993261e" containerID="0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75" exitCode=0 Dec 15 12:41:43 crc kubenswrapper[4719]: I1215 12:41:43.220650 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de640a25-d092-4c73-bb2f-c7a21993261e","Type":"ContainerDied","Data":"1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f"} Dec 15 12:41:43 crc kubenswrapper[4719]: I1215 12:41:43.220682 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de640a25-d092-4c73-bb2f-c7a21993261e","Type":"ContainerDied","Data":"2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8"} Dec 15 12:41:43 crc kubenswrapper[4719]: I1215 12:41:43.220696 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de640a25-d092-4c73-bb2f-c7a21993261e","Type":"ContainerDied","Data":"0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75"} Dec 15 12:41:43 crc kubenswrapper[4719]: I1215 12:41:43.223736 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" event={"ID":"2fd5121b-dd23-4935-a6d2-3a81e92193cf","Type":"ContainerStarted","Data":"cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685"} Dec 15 12:41:43 crc kubenswrapper[4719]: I1215 12:41:43.224040 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:43 crc kubenswrapper[4719]: I1215 12:41:43.244821 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" podStartSLOduration=5.244803762 podStartE2EDuration="5.244803762s" podCreationTimestamp="2025-12-15 12:41:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:41:43.239294749 +0000 UTC m=+1464.181587769" watchObservedRunningTime="2025-12-15 12:41:43.244803762 +0000 UTC m=+1464.187096792" Dec 15 12:41:44 crc kubenswrapper[4719]: I1215 12:41:44.873175 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:41:44 crc kubenswrapper[4719]: I1215 12:41:44.946622 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-combined-ca-bundle\") pod \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " Dec 15 12:41:44 crc kubenswrapper[4719]: I1215 12:41:44.946772 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzf7p\" (UniqueName: \"kubernetes.io/projected/572d65f8-13e4-4e29-b7e1-84d489b0acb0-kube-api-access-mzf7p\") pod \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " Dec 15 12:41:44 crc kubenswrapper[4719]: I1215 12:41:44.946818 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572d65f8-13e4-4e29-b7e1-84d489b0acb0-logs\") pod \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " Dec 15 12:41:44 crc kubenswrapper[4719]: I1215 12:41:44.947038 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-config-data\") pod \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\" (UID: \"572d65f8-13e4-4e29-b7e1-84d489b0acb0\") " Dec 15 12:41:44 crc kubenswrapper[4719]: I1215 12:41:44.950391 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572d65f8-13e4-4e29-b7e1-84d489b0acb0-logs" (OuterVolumeSpecName: "logs") pod "572d65f8-13e4-4e29-b7e1-84d489b0acb0" (UID: "572d65f8-13e4-4e29-b7e1-84d489b0acb0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:44 crc kubenswrapper[4719]: I1215 12:41:44.962597 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/572d65f8-13e4-4e29-b7e1-84d489b0acb0-kube-api-access-mzf7p" (OuterVolumeSpecName: "kube-api-access-mzf7p") pod "572d65f8-13e4-4e29-b7e1-84d489b0acb0" (UID: "572d65f8-13e4-4e29-b7e1-84d489b0acb0"). InnerVolumeSpecName "kube-api-access-mzf7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:44 crc kubenswrapper[4719]: I1215 12:41:44.981514 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:41:44 crc kubenswrapper[4719]: I1215 12:41:44.993740 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "572d65f8-13e4-4e29-b7e1-84d489b0acb0" (UID: "572d65f8-13e4-4e29-b7e1-84d489b0acb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:44 crc kubenswrapper[4719]: I1215 12:41:44.996794 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-config-data" (OuterVolumeSpecName: "config-data") pod "572d65f8-13e4-4e29-b7e1-84d489b0acb0" (UID: "572d65f8-13e4-4e29-b7e1-84d489b0acb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.048739 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-run-httpd\") pod \"de640a25-d092-4c73-bb2f-c7a21993261e\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.048813 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-log-httpd\") pod \"de640a25-d092-4c73-bb2f-c7a21993261e\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.048933 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-ceilometer-tls-certs\") pod \"de640a25-d092-4c73-bb2f-c7a21993261e\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.048957 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-config-data\") pod \"de640a25-d092-4c73-bb2f-c7a21993261e\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.048996 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-combined-ca-bundle\") pod \"de640a25-d092-4c73-bb2f-c7a21993261e\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.049022 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-scripts\") pod \"de640a25-d092-4c73-bb2f-c7a21993261e\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.049070 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crlrx\" (UniqueName: \"kubernetes.io/projected/de640a25-d092-4c73-bb2f-c7a21993261e-kube-api-access-crlrx\") pod \"de640a25-d092-4c73-bb2f-c7a21993261e\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.049163 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-sg-core-conf-yaml\") pod \"de640a25-d092-4c73-bb2f-c7a21993261e\" (UID: \"de640a25-d092-4c73-bb2f-c7a21993261e\") " Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.049653 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.049672 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/572d65f8-13e4-4e29-b7e1-84d489b0acb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.049683 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzf7p\" (UniqueName: \"kubernetes.io/projected/572d65f8-13e4-4e29-b7e1-84d489b0acb0-kube-api-access-mzf7p\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.049693 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/572d65f8-13e4-4e29-b7e1-84d489b0acb0-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.050642 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "de640a25-d092-4c73-bb2f-c7a21993261e" (UID: "de640a25-d092-4c73-bb2f-c7a21993261e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.050864 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "de640a25-d092-4c73-bb2f-c7a21993261e" (UID: "de640a25-d092-4c73-bb2f-c7a21993261e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.058417 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de640a25-d092-4c73-bb2f-c7a21993261e-kube-api-access-crlrx" (OuterVolumeSpecName: "kube-api-access-crlrx") pod "de640a25-d092-4c73-bb2f-c7a21993261e" (UID: "de640a25-d092-4c73-bb2f-c7a21993261e"). InnerVolumeSpecName "kube-api-access-crlrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.059120 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-scripts" (OuterVolumeSpecName: "scripts") pod "de640a25-d092-4c73-bb2f-c7a21993261e" (UID: "de640a25-d092-4c73-bb2f-c7a21993261e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.095269 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "de640a25-d092-4c73-bb2f-c7a21993261e" (UID: "de640a25-d092-4c73-bb2f-c7a21993261e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.122626 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "de640a25-d092-4c73-bb2f-c7a21993261e" (UID: "de640a25-d092-4c73-bb2f-c7a21993261e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.152117 4719 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.152169 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.152184 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crlrx\" (UniqueName: \"kubernetes.io/projected/de640a25-d092-4c73-bb2f-c7a21993261e-kube-api-access-crlrx\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.152228 4719 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.152239 4719 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.152249 4719 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de640a25-d092-4c73-bb2f-c7a21993261e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.162553 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de640a25-d092-4c73-bb2f-c7a21993261e" (UID: "de640a25-d092-4c73-bb2f-c7a21993261e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.182324 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-config-data" (OuterVolumeSpecName: "config-data") pod "de640a25-d092-4c73-bb2f-c7a21993261e" (UID: "de640a25-d092-4c73-bb2f-c7a21993261e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.241697 4719 generic.go:334] "Generic (PLEG): container finished" podID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerID="b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b" exitCode=0 Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.241900 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"572d65f8-13e4-4e29-b7e1-84d489b0acb0","Type":"ContainerDied","Data":"b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b"} Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.242008 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.242111 4719 scope.go:117] "RemoveContainer" containerID="b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.242040 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"572d65f8-13e4-4e29-b7e1-84d489b0acb0","Type":"ContainerDied","Data":"59f2423eda0c2a6ad7a4e71909ec2d5b619ce4d6d3f5a706f081f85a68b71d73"} Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.245671 4719 generic.go:334] "Generic (PLEG): container finished" podID="de640a25-d092-4c73-bb2f-c7a21993261e" containerID="d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54" exitCode=0 Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.245703 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de640a25-d092-4c73-bb2f-c7a21993261e","Type":"ContainerDied","Data":"d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54"} Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.245728 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de640a25-d092-4c73-bb2f-c7a21993261e","Type":"ContainerDied","Data":"073ac7817c37ab63faa1b3e17c81e375e655ad106a1f5542cb09228f2404384d"} Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.245783 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.253708 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.253735 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de640a25-d092-4c73-bb2f-c7a21993261e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.278146 4719 scope.go:117] "RemoveContainer" containerID="782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.303045 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.325131 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.341343 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.345891 4719 scope.go:117] "RemoveContainer" containerID="b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.347519 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b\": container with ID starting with b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b not found: ID does not exist" containerID="b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.347556 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b"} err="failed to get container status \"b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b\": rpc error: code = NotFound desc = could not find container \"b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b\": container with ID starting with b36cd4bf62f1e81f0f236fd16a9341beeccace1cd7eaf5763bc2a90bc01ea60b not found: ID does not exist" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.347587 4719 scope.go:117] "RemoveContainer" containerID="782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.348848 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528\": container with ID starting with 782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528 not found: ID does not exist" containerID="782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.348901 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528"} err="failed to get container status \"782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528\": rpc error: code = NotFound desc = could not find container \"782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528\": container with ID starting with 782772c514a165b2f9ea21e12ec47b968ec4474e762d59fbcf316d21ff13b528 not found: ID does not exist" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.348928 4719 scope.go:117] "RemoveContainer" containerID="1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.349946 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.380907 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.381399 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="sg-core" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381425 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="sg-core" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.381451 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="ceilometer-notification-agent" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381460 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="ceilometer-notification-agent" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.381477 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerName="nova-api-api" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381483 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerName="nova-api-api" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.381490 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerName="nova-api-log" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381496 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerName="nova-api-log" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.381514 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="ceilometer-central-agent" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381520 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="ceilometer-central-agent" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.381535 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="proxy-httpd" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381541 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="proxy-httpd" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381563 4719 scope.go:117] "RemoveContainer" containerID="2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381761 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="ceilometer-central-agent" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381777 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="ceilometer-notification-agent" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381791 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerName="nova-api-log" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381806 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" containerName="nova-api-api" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381825 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="sg-core" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.381841 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" containerName="proxy-httpd" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.382848 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.395559 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.395583 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.397231 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.436017 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.455307 4719 scope.go:117] "RemoveContainer" containerID="d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.464216 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.467037 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.469620 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.469620 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.517260 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="572d65f8-13e4-4e29-b7e1-84d489b0acb0" path="/var/lib/kubelet/pods/572d65f8-13e4-4e29-b7e1-84d489b0acb0/volumes" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.518173 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de640a25-d092-4c73-bb2f-c7a21993261e" path="/var/lib/kubelet/pods/de640a25-d092-4c73-bb2f-c7a21993261e/volumes" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.529954 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.529985 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.536035 4719 scope.go:117] "RemoveContainer" containerID="0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.566729 4719 scope.go:117] "RemoveContainer" containerID="1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.567080 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f\": container with ID starting with 1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f not found: ID does not exist" containerID="1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567107 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f"} err="failed to get container status \"1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f\": rpc error: code = NotFound desc = could not find container \"1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f\": container with ID starting with 1846e1ca5a818a6e2cebc4a6d62c2fd335e27e28153d7a2b62c7edc2e724531f not found: ID does not exist" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567127 4719 scope.go:117] "RemoveContainer" containerID="2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567321 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz2zn\" (UniqueName: \"kubernetes.io/projected/0a4f5fc1-37ce-4a26-8693-23776cf567cb-kube-api-access-dz2zn\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.567344 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8\": container with ID starting with 2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8 not found: ID does not exist" containerID="2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567398 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-log-httpd\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567408 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8"} err="failed to get container status \"2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8\": rpc error: code = NotFound desc = could not find container \"2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8\": container with ID starting with 2cd6715664b14f17470f815ff81415ef0127a5c3ad74aa01c61463bbf323e3a8 not found: ID does not exist" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567426 4719 scope.go:117] "RemoveContainer" containerID="d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567429 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-run-httpd\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567463 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567568 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4f5fc1-37ce-4a26-8693-23776cf567cb-logs\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567613 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567638 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.567654 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54\": container with ID starting with d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54 not found: ID does not exist" containerID="d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567670 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54"} err="failed to get container status \"d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54\": rpc error: code = NotFound desc = could not find container \"d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54\": container with ID starting with d83078d24a5cdb6b11a6039ad046a50757511060abd3eca8fc227fcd62583e54 not found: ID does not exist" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567694 4719 scope.go:117] "RemoveContainer" containerID="0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567693 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567731 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-scripts\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567766 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567806 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-config-data\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567851 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567895 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wmbd\" (UniqueName: \"kubernetes.io/projected/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-kube-api-access-2wmbd\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567926 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-config-data\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: E1215 12:41:45.567941 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75\": container with ID starting with 0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75 not found: ID does not exist" containerID="0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.567981 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75"} err="failed to get container status \"0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75\": rpc error: code = NotFound desc = could not find container \"0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75\": container with ID starting with 0baa6102095ff17b692f5fa520fb92410bef0e57084e97cd7d1c8488b2c29e75 not found: ID does not exist" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669190 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-config-data\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669308 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz2zn\" (UniqueName: \"kubernetes.io/projected/0a4f5fc1-37ce-4a26-8693-23776cf567cb-kube-api-access-dz2zn\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669350 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-log-httpd\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669375 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-run-httpd\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669418 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669470 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4f5fc1-37ce-4a26-8693-23776cf567cb-logs\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669511 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669537 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669577 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669624 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-scripts\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669657 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669702 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-config-data\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669737 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.669759 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wmbd\" (UniqueName: \"kubernetes.io/projected/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-kube-api-access-2wmbd\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.671820 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-run-httpd\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.671883 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-log-httpd\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.676075 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.676812 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4f5fc1-37ce-4a26-8693-23776cf567cb-logs\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.677708 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.678728 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.678957 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-config-data\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.680396 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.680977 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.681240 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.686270 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-config-data\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.686460 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-scripts\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.688357 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wmbd\" (UniqueName: \"kubernetes.io/projected/25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0-kube-api-access-2wmbd\") pod \"ceilometer-0\" (UID: \"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0\") " pod="openstack/ceilometer-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.688363 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz2zn\" (UniqueName: \"kubernetes.io/projected/0a4f5fc1-37ce-4a26-8693-23776cf567cb-kube-api-access-dz2zn\") pod \"nova-api-0\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.835384 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:41:45 crc kubenswrapper[4719]: I1215 12:41:45.848774 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:46.818920 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:46.850608 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:47.279246 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:48.047031 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-zkv6s" podUID="e6adc1e0-5c73-431d-a94d-e2eabc16dd03" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.78:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:48.807647 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-j5gcb"] Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:48.808699 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:48.812937 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:48.812974 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:48.819678 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-j5gcb"] Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:48.931026 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmxtz\" (UniqueName: \"kubernetes.io/projected/33983783-3b6b-4a9e-93c0-2cf842823bf3-kube-api-access-tmxtz\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:48.931077 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-scripts\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:48.931133 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:48.931324 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-config-data\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.019046 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.034997 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-config-data\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.035098 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmxtz\" (UniqueName: \"kubernetes.io/projected/33983783-3b6b-4a9e-93c0-2cf842823bf3-kube-api-access-tmxtz\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.035131 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-scripts\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.035181 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.065747 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-config-data\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.068338 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-scripts\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.081575 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.096365 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmxtz\" (UniqueName: \"kubernetes.io/projected/33983783-3b6b-4a9e-93c0-2cf842823bf3-kube-api-access-tmxtz\") pod \"nova-cell1-cell-mapping-j5gcb\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.139746 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.178500 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-jd2vn"] Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.178806 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" podUID="e8230fff-add6-4db5-b288-638524c1d197" containerName="dnsmasq-dns" containerID="cri-o://479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe" gracePeriod=10 Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.609691 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.813954 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 15 12:41:49 crc kubenswrapper[4719]: I1215 12:41:49.936108 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-j5gcb"] Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.297009 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.325637 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4f5fc1-37ce-4a26-8693-23776cf567cb","Type":"ContainerStarted","Data":"3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5"} Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.325683 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4f5fc1-37ce-4a26-8693-23776cf567cb","Type":"ContainerStarted","Data":"e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c"} Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.325693 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4f5fc1-37ce-4a26-8693-23776cf567cb","Type":"ContainerStarted","Data":"c88ed89cd0d5c3adb20c26e6a76dccac984565a465c9520e352a07faa4689d27"} Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.327876 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0","Type":"ContainerStarted","Data":"b1ee9374d1a759c8c9d1fad8a3309be223dfb9a16ba4b4357b68c701db84034e"} Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.330934 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-j5gcb" event={"ID":"33983783-3b6b-4a9e-93c0-2cf842823bf3","Type":"ContainerStarted","Data":"efed273c49a2a66b71d08f0e27fe9936e76f07705d036b6cac06b15eca91b310"} Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.337498 4719 generic.go:334] "Generic (PLEG): container finished" podID="e8230fff-add6-4db5-b288-638524c1d197" containerID="479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe" exitCode=0 Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.337538 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" event={"ID":"e8230fff-add6-4db5-b288-638524c1d197","Type":"ContainerDied","Data":"479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe"} Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.337562 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" event={"ID":"e8230fff-add6-4db5-b288-638524c1d197","Type":"ContainerDied","Data":"1b4a54ebb2c4089f6df324830b4820cd9b6048ef58d22a8bfb670bb4021f7453"} Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.337581 4719 scope.go:117] "RemoveContainer" containerID="479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.337695 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-jd2vn" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.364342 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=5.364322002 podStartE2EDuration="5.364322002s" podCreationTimestamp="2025-12-15 12:41:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:41:50.352741778 +0000 UTC m=+1471.295034808" watchObservedRunningTime="2025-12-15 12:41:50.364322002 +0000 UTC m=+1471.306615032" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.369597 4719 scope.go:117] "RemoveContainer" containerID="e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.439564 4719 scope.go:117] "RemoveContainer" containerID="479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe" Dec 15 12:41:50 crc kubenswrapper[4719]: E1215 12:41:50.440469 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe\": container with ID starting with 479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe not found: ID does not exist" containerID="479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.440550 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe"} err="failed to get container status \"479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe\": rpc error: code = NotFound desc = could not find container \"479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe\": container with ID starting with 479abfe592d93152105fcbebbd61e25eeb6ba87b05a9dd3f8335e598d28836fe not found: ID does not exist" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.440652 4719 scope.go:117] "RemoveContainer" containerID="e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10" Dec 15 12:41:50 crc kubenswrapper[4719]: E1215 12:41:50.446517 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10\": container with ID starting with e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10 not found: ID does not exist" containerID="e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.446561 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10"} err="failed to get container status \"e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10\": rpc error: code = NotFound desc = could not find container \"e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10\": container with ID starting with e44481ac6b3b667452205d537ba4253aec0f5f18202039a98e03d75aa1b98c10 not found: ID does not exist" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.472469 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-sb\") pod \"e8230fff-add6-4db5-b288-638524c1d197\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.472528 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9c4f\" (UniqueName: \"kubernetes.io/projected/e8230fff-add6-4db5-b288-638524c1d197-kube-api-access-n9c4f\") pod \"e8230fff-add6-4db5-b288-638524c1d197\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.472580 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-nb\") pod \"e8230fff-add6-4db5-b288-638524c1d197\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.472659 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-svc\") pod \"e8230fff-add6-4db5-b288-638524c1d197\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.472700 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-swift-storage-0\") pod \"e8230fff-add6-4db5-b288-638524c1d197\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.472714 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-config\") pod \"e8230fff-add6-4db5-b288-638524c1d197\" (UID: \"e8230fff-add6-4db5-b288-638524c1d197\") " Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.505784 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8230fff-add6-4db5-b288-638524c1d197-kube-api-access-n9c4f" (OuterVolumeSpecName: "kube-api-access-n9c4f") pod "e8230fff-add6-4db5-b288-638524c1d197" (UID: "e8230fff-add6-4db5-b288-638524c1d197"). InnerVolumeSpecName "kube-api-access-n9c4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.556232 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e8230fff-add6-4db5-b288-638524c1d197" (UID: "e8230fff-add6-4db5-b288-638524c1d197"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.557024 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e8230fff-add6-4db5-b288-638524c1d197" (UID: "e8230fff-add6-4db5-b288-638524c1d197"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.566114 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8230fff-add6-4db5-b288-638524c1d197" (UID: "e8230fff-add6-4db5-b288-638524c1d197"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.575066 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.575960 4719 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.576000 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.576012 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9c4f\" (UniqueName: \"kubernetes.io/projected/e8230fff-add6-4db5-b288-638524c1d197-kube-api-access-n9c4f\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.590239 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e8230fff-add6-4db5-b288-638524c1d197" (UID: "e8230fff-add6-4db5-b288-638524c1d197"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.592953 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-config" (OuterVolumeSpecName: "config") pod "e8230fff-add6-4db5-b288-638524c1d197" (UID: "e8230fff-add6-4db5-b288-638524c1d197"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.668509 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-jd2vn"] Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.679123 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-jd2vn"] Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.680423 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:50 crc kubenswrapper[4719]: I1215 12:41:50.680445 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8230fff-add6-4db5-b288-638524c1d197-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:51 crc kubenswrapper[4719]: I1215 12:41:51.358666 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0","Type":"ContainerStarted","Data":"1bbbe050626962e2af83c86499150a5e256758f56c5ef5d14b83bae0cfee24ac"} Dec 15 12:41:51 crc kubenswrapper[4719]: I1215 12:41:51.372982 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-j5gcb" event={"ID":"33983783-3b6b-4a9e-93c0-2cf842823bf3","Type":"ContainerStarted","Data":"e98e660ab24653045210f1db1ceaca509ac69edeb24cfadc11978986dc3e8fc2"} Dec 15 12:41:51 crc kubenswrapper[4719]: I1215 12:41:51.397841 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-j5gcb" podStartSLOduration=3.397822456 podStartE2EDuration="3.397822456s" podCreationTimestamp="2025-12-15 12:41:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:41:51.389117883 +0000 UTC m=+1472.331410913" watchObservedRunningTime="2025-12-15 12:41:51.397822456 +0000 UTC m=+1472.340115486" Dec 15 12:41:51 crc kubenswrapper[4719]: I1215 12:41:51.487887 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8230fff-add6-4db5-b288-638524c1d197" path="/var/lib/kubelet/pods/e8230fff-add6-4db5-b288-638524c1d197/volumes" Dec 15 12:41:52 crc kubenswrapper[4719]: I1215 12:41:52.385213 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0","Type":"ContainerStarted","Data":"47e6843614e65a7de5fce168e1b4dad1199c46807491fabca4953e5464282657"} Dec 15 12:41:53 crc kubenswrapper[4719]: I1215 12:41:53.394072 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0","Type":"ContainerStarted","Data":"b1836ab86a502524a53530123a07c5b3f6f769a25bdfc76bf306a658daee49a9"} Dec 15 12:41:55 crc kubenswrapper[4719]: I1215 12:41:55.835603 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 15 12:41:55 crc kubenswrapper[4719]: I1215 12:41:55.836194 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 15 12:41:56 crc kubenswrapper[4719]: I1215 12:41:56.421252 4719 generic.go:334] "Generic (PLEG): container finished" podID="33983783-3b6b-4a9e-93c0-2cf842823bf3" containerID="e98e660ab24653045210f1db1ceaca509ac69edeb24cfadc11978986dc3e8fc2" exitCode=0 Dec 15 12:41:56 crc kubenswrapper[4719]: I1215 12:41:56.421274 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-j5gcb" event={"ID":"33983783-3b6b-4a9e-93c0-2cf842823bf3","Type":"ContainerDied","Data":"e98e660ab24653045210f1db1ceaca509ac69edeb24cfadc11978986dc3e8fc2"} Dec 15 12:41:56 crc kubenswrapper[4719]: I1215 12:41:56.882116 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:41:56 crc kubenswrapper[4719]: I1215 12:41:56.882224 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.076543 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.232502 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-config-data\") pod \"33983783-3b6b-4a9e-93c0-2cf842823bf3\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.232605 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-scripts\") pod \"33983783-3b6b-4a9e-93c0-2cf842823bf3\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.232682 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-combined-ca-bundle\") pod \"33983783-3b6b-4a9e-93c0-2cf842823bf3\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.232749 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmxtz\" (UniqueName: \"kubernetes.io/projected/33983783-3b6b-4a9e-93c0-2cf842823bf3-kube-api-access-tmxtz\") pod \"33983783-3b6b-4a9e-93c0-2cf842823bf3\" (UID: \"33983783-3b6b-4a9e-93c0-2cf842823bf3\") " Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.238159 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-scripts" (OuterVolumeSpecName: "scripts") pod "33983783-3b6b-4a9e-93c0-2cf842823bf3" (UID: "33983783-3b6b-4a9e-93c0-2cf842823bf3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.238978 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33983783-3b6b-4a9e-93c0-2cf842823bf3-kube-api-access-tmxtz" (OuterVolumeSpecName: "kube-api-access-tmxtz") pod "33983783-3b6b-4a9e-93c0-2cf842823bf3" (UID: "33983783-3b6b-4a9e-93c0-2cf842823bf3"). InnerVolumeSpecName "kube-api-access-tmxtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.265396 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-config-data" (OuterVolumeSpecName: "config-data") pod "33983783-3b6b-4a9e-93c0-2cf842823bf3" (UID: "33983783-3b6b-4a9e-93c0-2cf842823bf3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.268036 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33983783-3b6b-4a9e-93c0-2cf842823bf3" (UID: "33983783-3b6b-4a9e-93c0-2cf842823bf3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.334437 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.334475 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmxtz\" (UniqueName: \"kubernetes.io/projected/33983783-3b6b-4a9e-93c0-2cf842823bf3-kube-api-access-tmxtz\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.334487 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.334495 4719 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33983783-3b6b-4a9e-93c0-2cf842823bf3-scripts\") on node \"crc\" DevicePath \"\"" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.455772 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0","Type":"ContainerStarted","Data":"c1a53dcc4d3e3bbe22886b64193fcdc1eb30a2bb49a4d383ee7a37c16fc2a9a2"} Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.457387 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.460123 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-j5gcb" event={"ID":"33983783-3b6b-4a9e-93c0-2cf842823bf3","Type":"ContainerDied","Data":"efed273c49a2a66b71d08f0e27fe9936e76f07705d036b6cac06b15eca91b310"} Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.460160 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efed273c49a2a66b71d08f0e27fe9936e76f07705d036b6cac06b15eca91b310" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.460194 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-j5gcb" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.628076 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.6786028680000005 podStartE2EDuration="13.628053096s" podCreationTimestamp="2025-12-15 12:41:45 +0000 UTC" firstStartedPulling="2025-12-15 12:41:49.820345839 +0000 UTC m=+1470.762638869" lastFinishedPulling="2025-12-15 12:41:57.769796067 +0000 UTC m=+1478.712089097" observedRunningTime="2025-12-15 12:41:58.486624342 +0000 UTC m=+1479.428917382" watchObservedRunningTime="2025-12-15 12:41:58.628053096 +0000 UTC m=+1479.570346126" Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.638679 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.639175 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="be6799a7-1506-4f57-abba-a0db463847c9" containerName="nova-scheduler-scheduler" containerID="cri-o://17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625" gracePeriod=30 Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.650595 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.650922 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-log" containerID="cri-o://e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be" gracePeriod=30 Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.651066 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-metadata" containerID="cri-o://c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637" gracePeriod=30 Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.664987 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.665506 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerName="nova-api-log" containerID="cri-o://e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c" gracePeriod=30 Dec 15 12:41:58 crc kubenswrapper[4719]: I1215 12:41:58.666087 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerName="nova-api-api" containerID="cri-o://3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5" gracePeriod=30 Dec 15 12:41:59 crc kubenswrapper[4719]: I1215 12:41:59.484475 4719 generic.go:334] "Generic (PLEG): container finished" podID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerID="e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be" exitCode=143 Dec 15 12:41:59 crc kubenswrapper[4719]: I1215 12:41:59.487332 4719 generic.go:334] "Generic (PLEG): container finished" podID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerID="e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c" exitCode=143 Dec 15 12:41:59 crc kubenswrapper[4719]: I1215 12:41:59.489616 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80","Type":"ContainerDied","Data":"e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be"} Dec 15 12:41:59 crc kubenswrapper[4719]: I1215 12:41:59.489691 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4f5fc1-37ce-4a26-8693-23776cf567cb","Type":"ContainerDied","Data":"e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c"} Dec 15 12:42:00 crc kubenswrapper[4719]: E1215 12:42:00.491268 4719 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 15 12:42:00 crc kubenswrapper[4719]: E1215 12:42:00.494444 4719 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 15 12:42:00 crc kubenswrapper[4719]: E1215 12:42:00.496666 4719 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 15 12:42:00 crc kubenswrapper[4719]: E1215 12:42:00.496767 4719 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="be6799a7-1506-4f57-abba-a0db463847c9" containerName="nova-scheduler-scheduler" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.355259 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.514194 4719 generic.go:334] "Generic (PLEG): container finished" podID="be6799a7-1506-4f57-abba-a0db463847c9" containerID="17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625" exitCode=0 Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.514246 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be6799a7-1506-4f57-abba-a0db463847c9","Type":"ContainerDied","Data":"17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625"} Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.514280 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be6799a7-1506-4f57-abba-a0db463847c9","Type":"ContainerDied","Data":"c93b14720c53b973f522fbca48014e3e3894a38553fcaa0d853e0d6adbb90741"} Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.514300 4719 scope.go:117] "RemoveContainer" containerID="17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.514478 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.516796 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xmld\" (UniqueName: \"kubernetes.io/projected/be6799a7-1506-4f57-abba-a0db463847c9-kube-api-access-6xmld\") pod \"be6799a7-1506-4f57-abba-a0db463847c9\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.516990 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-combined-ca-bundle\") pod \"be6799a7-1506-4f57-abba-a0db463847c9\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.517071 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-config-data\") pod \"be6799a7-1506-4f57-abba-a0db463847c9\" (UID: \"be6799a7-1506-4f57-abba-a0db463847c9\") " Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.535204 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be6799a7-1506-4f57-abba-a0db463847c9-kube-api-access-6xmld" (OuterVolumeSpecName: "kube-api-access-6xmld") pod "be6799a7-1506-4f57-abba-a0db463847c9" (UID: "be6799a7-1506-4f57-abba-a0db463847c9"). InnerVolumeSpecName "kube-api-access-6xmld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.548825 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be6799a7-1506-4f57-abba-a0db463847c9" (UID: "be6799a7-1506-4f57-abba-a0db463847c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.549205 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-config-data" (OuterVolumeSpecName: "config-data") pod "be6799a7-1506-4f57-abba-a0db463847c9" (UID: "be6799a7-1506-4f57-abba-a0db463847c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.618703 4719 scope.go:117] "RemoveContainer" containerID="17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.618916 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.618933 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be6799a7-1506-4f57-abba-a0db463847c9-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.618942 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xmld\" (UniqueName: \"kubernetes.io/projected/be6799a7-1506-4f57-abba-a0db463847c9-kube-api-access-6xmld\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:01 crc kubenswrapper[4719]: E1215 12:42:01.619331 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625\": container with ID starting with 17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625 not found: ID does not exist" containerID="17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.619362 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625"} err="failed to get container status \"17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625\": rpc error: code = NotFound desc = could not find container \"17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625\": container with ID starting with 17508cb9fdf7ac03d31768c45c8dd9c66065a24ea7b178da617cd52facfde625 not found: ID does not exist" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.784037 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": read tcp 10.217.0.2:50704->10.217.0.196:8775: read: connection reset by peer" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.784142 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": read tcp 10.217.0.2:50714->10.217.0.196:8775: read: connection reset by peer" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.949535 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.958016 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.977360 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:42:01 crc kubenswrapper[4719]: E1215 12:42:01.977696 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8230fff-add6-4db5-b288-638524c1d197" containerName="dnsmasq-dns" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.977707 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8230fff-add6-4db5-b288-638524c1d197" containerName="dnsmasq-dns" Dec 15 12:42:01 crc kubenswrapper[4719]: E1215 12:42:01.977717 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8230fff-add6-4db5-b288-638524c1d197" containerName="init" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.977723 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8230fff-add6-4db5-b288-638524c1d197" containerName="init" Dec 15 12:42:01 crc kubenswrapper[4719]: E1215 12:42:01.977744 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33983783-3b6b-4a9e-93c0-2cf842823bf3" containerName="nova-manage" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.977749 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="33983783-3b6b-4a9e-93c0-2cf842823bf3" containerName="nova-manage" Dec 15 12:42:01 crc kubenswrapper[4719]: E1215 12:42:01.977759 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be6799a7-1506-4f57-abba-a0db463847c9" containerName="nova-scheduler-scheduler" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.977765 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="be6799a7-1506-4f57-abba-a0db463847c9" containerName="nova-scheduler-scheduler" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.977945 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="be6799a7-1506-4f57-abba-a0db463847c9" containerName="nova-scheduler-scheduler" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.977960 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8230fff-add6-4db5-b288-638524c1d197" containerName="dnsmasq-dns" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.977973 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="33983783-3b6b-4a9e-93c0-2cf842823bf3" containerName="nova-manage" Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.978530 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:42:01 crc kubenswrapper[4719]: I1215 12:42:01.978603 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.005366 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.127349 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0945867e-d6c4-4f3e-976d-b2ae68497a11-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0945867e-d6c4-4f3e-976d-b2ae68497a11\") " pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.127843 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kxpg\" (UniqueName: \"kubernetes.io/projected/0945867e-d6c4-4f3e-976d-b2ae68497a11-kube-api-access-8kxpg\") pod \"nova-scheduler-0\" (UID: \"0945867e-d6c4-4f3e-976d-b2ae68497a11\") " pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.128005 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0945867e-d6c4-4f3e-976d-b2ae68497a11-config-data\") pod \"nova-scheduler-0\" (UID: \"0945867e-d6c4-4f3e-976d-b2ae68497a11\") " pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.230288 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0945867e-d6c4-4f3e-976d-b2ae68497a11-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0945867e-d6c4-4f3e-976d-b2ae68497a11\") " pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.230360 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kxpg\" (UniqueName: \"kubernetes.io/projected/0945867e-d6c4-4f3e-976d-b2ae68497a11-kube-api-access-8kxpg\") pod \"nova-scheduler-0\" (UID: \"0945867e-d6c4-4f3e-976d-b2ae68497a11\") " pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.230681 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0945867e-d6c4-4f3e-976d-b2ae68497a11-config-data\") pod \"nova-scheduler-0\" (UID: \"0945867e-d6c4-4f3e-976d-b2ae68497a11\") " pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.235363 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0945867e-d6c4-4f3e-976d-b2ae68497a11-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0945867e-d6c4-4f3e-976d-b2ae68497a11\") " pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.250421 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0945867e-d6c4-4f3e-976d-b2ae68497a11-config-data\") pod \"nova-scheduler-0\" (UID: \"0945867e-d6c4-4f3e-976d-b2ae68497a11\") " pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.260638 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kxpg\" (UniqueName: \"kubernetes.io/projected/0945867e-d6c4-4f3e-976d-b2ae68497a11-kube-api-access-8kxpg\") pod \"nova-scheduler-0\" (UID: \"0945867e-d6c4-4f3e-976d-b2ae68497a11\") " pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.364882 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.369461 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.379298 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.534740 4719 generic.go:334] "Generic (PLEG): container finished" podID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerID="c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637" exitCode=0 Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.534884 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.534964 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80","Type":"ContainerDied","Data":"c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637"} Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.535236 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80","Type":"ContainerDied","Data":"6c5d65692b93dd1a96b94a112de2a4349eabbeccd94613abfe2278e76ccb0c3d"} Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.535257 4719 scope.go:117] "RemoveContainer" containerID="c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.539809 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz2zn\" (UniqueName: \"kubernetes.io/projected/0a4f5fc1-37ce-4a26-8693-23776cf567cb-kube-api-access-dz2zn\") pod \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.543972 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4f5fc1-37ce-4a26-8693-23776cf567cb-logs\") pod \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.544131 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-public-tls-certs\") pod \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.544218 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-config-data\") pod \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.544307 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-config-data\") pod \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.544314 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a4f5fc1-37ce-4a26-8693-23776cf567cb-logs" (OuterVolumeSpecName: "logs") pod "0a4f5fc1-37ce-4a26-8693-23776cf567cb" (UID: "0a4f5fc1-37ce-4a26-8693-23776cf567cb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.544401 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-combined-ca-bundle\") pod \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.544481 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-nova-metadata-tls-certs\") pod \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.544552 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zrn5\" (UniqueName: \"kubernetes.io/projected/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-kube-api-access-9zrn5\") pod \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.544616 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-logs\") pod \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\" (UID: \"22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.544652 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-combined-ca-bundle\") pod \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.544707 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-internal-tls-certs\") pod \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\" (UID: \"0a4f5fc1-37ce-4a26-8693-23776cf567cb\") " Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.547115 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4f5fc1-37ce-4a26-8693-23776cf567cb-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.549385 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-logs" (OuterVolumeSpecName: "logs") pod "22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" (UID: "22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.554532 4719 generic.go:334] "Generic (PLEG): container finished" podID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerID="3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5" exitCode=0 Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.554584 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4f5fc1-37ce-4a26-8693-23776cf567cb","Type":"ContainerDied","Data":"3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5"} Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.554608 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4f5fc1-37ce-4a26-8693-23776cf567cb","Type":"ContainerDied","Data":"c88ed89cd0d5c3adb20c26e6a76dccac984565a465c9520e352a07faa4689d27"} Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.554723 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.565033 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-kube-api-access-9zrn5" (OuterVolumeSpecName: "kube-api-access-9zrn5") pod "22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" (UID: "22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80"). InnerVolumeSpecName "kube-api-access-9zrn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.565347 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a4f5fc1-37ce-4a26-8693-23776cf567cb-kube-api-access-dz2zn" (OuterVolumeSpecName: "kube-api-access-dz2zn") pod "0a4f5fc1-37ce-4a26-8693-23776cf567cb" (UID: "0a4f5fc1-37ce-4a26-8693-23776cf567cb"). InnerVolumeSpecName "kube-api-access-dz2zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.595742 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-config-data" (OuterVolumeSpecName: "config-data") pod "0a4f5fc1-37ce-4a26-8693-23776cf567cb" (UID: "0a4f5fc1-37ce-4a26-8693-23776cf567cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.596096 4719 scope.go:117] "RemoveContainer" containerID="e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.607200 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" (UID: "22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.610037 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a4f5fc1-37ce-4a26-8693-23776cf567cb" (UID: "0a4f5fc1-37ce-4a26-8693-23776cf567cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.622037 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-config-data" (OuterVolumeSpecName: "config-data") pod "22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" (UID: "22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.629572 4719 scope.go:117] "RemoveContainer" containerID="c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637" Dec 15 12:42:02 crc kubenswrapper[4719]: E1215 12:42:02.637282 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637\": container with ID starting with c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637 not found: ID does not exist" containerID="c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.637332 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637"} err="failed to get container status \"c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637\": rpc error: code = NotFound desc = could not find container \"c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637\": container with ID starting with c446056a797cbf5dc75ef065741d9ab30e644bc656e3970d6b76a73849455637 not found: ID does not exist" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.637364 4719 scope.go:117] "RemoveContainer" containerID="e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be" Dec 15 12:42:02 crc kubenswrapper[4719]: E1215 12:42:02.637940 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be\": container with ID starting with e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be not found: ID does not exist" containerID="e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.637964 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be"} err="failed to get container status \"e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be\": rpc error: code = NotFound desc = could not find container \"e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be\": container with ID starting with e3318e849f2750d789f36a93d179fbe290579285abbe8f284daf1ed69ccc19be not found: ID does not exist" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.637981 4719 scope.go:117] "RemoveContainer" containerID="3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.648883 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zrn5\" (UniqueName: \"kubernetes.io/projected/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-kube-api-access-9zrn5\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.648901 4719 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-logs\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.648912 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.648922 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz2zn\" (UniqueName: \"kubernetes.io/projected/0a4f5fc1-37ce-4a26-8693-23776cf567cb-kube-api-access-dz2zn\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.648931 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.648939 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.648946 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.655993 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0a4f5fc1-37ce-4a26-8693-23776cf567cb" (UID: "0a4f5fc1-37ce-4a26-8693-23776cf567cb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.671613 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" (UID: "22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.680152 4719 scope.go:117] "RemoveContainer" containerID="e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.694107 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0a4f5fc1-37ce-4a26-8693-23776cf567cb" (UID: "0a4f5fc1-37ce-4a26-8693-23776cf567cb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.705702 4719 scope.go:117] "RemoveContainer" containerID="3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5" Dec 15 12:42:02 crc kubenswrapper[4719]: E1215 12:42:02.706389 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5\": container with ID starting with 3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5 not found: ID does not exist" containerID="3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.706426 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5"} err="failed to get container status \"3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5\": rpc error: code = NotFound desc = could not find container \"3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5\": container with ID starting with 3be3110d270245fad9e2566c81abc95bb678bd4cf173b6c97b413fc67f26fea5 not found: ID does not exist" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.706455 4719 scope.go:117] "RemoveContainer" containerID="e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c" Dec 15 12:42:02 crc kubenswrapper[4719]: E1215 12:42:02.707024 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c\": container with ID starting with e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c not found: ID does not exist" containerID="e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.707047 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c"} err="failed to get container status \"e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c\": rpc error: code = NotFound desc = could not find container \"e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c\": container with ID starting with e33187d51884ccc81b849e0b8cc6c0701227de06c07fc0a07b911b5eb173989c not found: ID does not exist" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.751167 4719 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.751206 4719 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.751221 4719 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a4f5fc1-37ce-4a26-8693-23776cf567cb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.874132 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.886304 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.902889 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:42:02 crc kubenswrapper[4719]: E1215 12:42:02.903427 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerName="nova-api-log" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.903493 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerName="nova-api-log" Dec 15 12:42:02 crc kubenswrapper[4719]: E1215 12:42:02.903577 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerName="nova-api-api" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.903625 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerName="nova-api-api" Dec 15 12:42:02 crc kubenswrapper[4719]: E1215 12:42:02.903686 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-metadata" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.903732 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-metadata" Dec 15 12:42:02 crc kubenswrapper[4719]: E1215 12:42:02.903787 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-log" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.903834 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-log" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.904087 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-log" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.904166 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" containerName="nova-metadata-metadata" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.904221 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerName="nova-api-log" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.904293 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" containerName="nova-api-api" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.905266 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.907722 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.907969 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.924359 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.932418 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.940431 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.952133 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.970340 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.970449 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.974135 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.974627 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.974803 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 15 12:42:02 crc kubenswrapper[4719]: I1215 12:42:02.978158 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.062373 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-config-data\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.063022 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.063118 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.063244 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-logs\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.063349 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr6km\" (UniqueName: \"kubernetes.io/projected/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-kube-api-access-sr6km\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164475 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr6km\" (UniqueName: \"kubernetes.io/projected/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-kube-api-access-sr6km\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164538 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m7w7\" (UniqueName: \"kubernetes.io/projected/59d59387-003d-47ae-a322-48787db59367-kube-api-access-9m7w7\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164613 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-internal-tls-certs\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164644 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-config-data\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164669 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164703 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164726 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164745 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d59387-003d-47ae-a322-48787db59367-logs\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164765 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-public-tls-certs\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164789 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-config-data\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.164809 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-logs\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.165274 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-logs\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.170682 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-config-data\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.170702 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.179600 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.183436 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr6km\" (UniqueName: \"kubernetes.io/projected/45c8d521-b7f2-4f93-a6e8-94d6aaf85390-kube-api-access-sr6km\") pod \"nova-metadata-0\" (UID: \"45c8d521-b7f2-4f93-a6e8-94d6aaf85390\") " pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.232125 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.266190 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m7w7\" (UniqueName: \"kubernetes.io/projected/59d59387-003d-47ae-a322-48787db59367-kube-api-access-9m7w7\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.266265 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-internal-tls-certs\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.266347 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.266372 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d59387-003d-47ae-a322-48787db59367-logs\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.266400 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-public-tls-certs\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.266427 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-config-data\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.267966 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d59387-003d-47ae-a322-48787db59367-logs\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.270899 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.271843 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-config-data\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.272786 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-public-tls-certs\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.273369 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d59387-003d-47ae-a322-48787db59367-internal-tls-certs\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.293451 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m7w7\" (UniqueName: \"kubernetes.io/projected/59d59387-003d-47ae-a322-48787db59367-kube-api-access-9m7w7\") pod \"nova-api-0\" (UID: \"59d59387-003d-47ae-a322-48787db59367\") " pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.430703 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.487269 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a4f5fc1-37ce-4a26-8693-23776cf567cb" path="/var/lib/kubelet/pods/0a4f5fc1-37ce-4a26-8693-23776cf567cb/volumes" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.488091 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80" path="/var/lib/kubelet/pods/22ba9cf5-ac33-4fa7-b23f-d0ad76f44f80/volumes" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.496403 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be6799a7-1506-4f57-abba-a0db463847c9" path="/var/lib/kubelet/pods/be6799a7-1506-4f57-abba-a0db463847c9/volumes" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.577842 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0945867e-d6c4-4f3e-976d-b2ae68497a11","Type":"ContainerStarted","Data":"7e9c5b2bbdd558769002a39e57369c5ca4cd377e7d84d12dc66860ea232c54b8"} Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.577921 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0945867e-d6c4-4f3e-976d-b2ae68497a11","Type":"ContainerStarted","Data":"1475b60573c26cd1144a70ca9880537fea1bf12c67cc9c1cd0d97e6e78a27d4f"} Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.608154 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.608133001 podStartE2EDuration="2.608133001s" podCreationTimestamp="2025-12-15 12:42:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:42:03.602182424 +0000 UTC m=+1484.544475474" watchObservedRunningTime="2025-12-15 12:42:03.608133001 +0000 UTC m=+1484.550426021" Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.684810 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 15 12:42:03 crc kubenswrapper[4719]: W1215 12:42:03.881987 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59d59387_003d_47ae_a322_48787db59367.slice/crio-8ae570c79702c6a36ba531e62a4133940c43e99c61aa559bd1bc56f2841224c8 WatchSource:0}: Error finding container 8ae570c79702c6a36ba531e62a4133940c43e99c61aa559bd1bc56f2841224c8: Status 404 returned error can't find the container with id 8ae570c79702c6a36ba531e62a4133940c43e99c61aa559bd1bc56f2841224c8 Dec 15 12:42:03 crc kubenswrapper[4719]: I1215 12:42:03.882320 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 15 12:42:04 crc kubenswrapper[4719]: I1215 12:42:04.612217 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59d59387-003d-47ae-a322-48787db59367","Type":"ContainerStarted","Data":"05905c49dcfedd0ed47360a0c0e919a6786d8c641d622d81df47aa9a39ca0f26"} Dec 15 12:42:04 crc kubenswrapper[4719]: I1215 12:42:04.614232 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59d59387-003d-47ae-a322-48787db59367","Type":"ContainerStarted","Data":"8ae570c79702c6a36ba531e62a4133940c43e99c61aa559bd1bc56f2841224c8"} Dec 15 12:42:04 crc kubenswrapper[4719]: I1215 12:42:04.625234 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"45c8d521-b7f2-4f93-a6e8-94d6aaf85390","Type":"ContainerStarted","Data":"43d691633714d9d56ac9e008edca8aa68ac63ba1a9c8b03522b5617a84a775f5"} Dec 15 12:42:04 crc kubenswrapper[4719]: I1215 12:42:04.625289 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"45c8d521-b7f2-4f93-a6e8-94d6aaf85390","Type":"ContainerStarted","Data":"8e17e497f3e9fca3a966acd7aee4d20c724004340d058f8d516c72f82ce30391"} Dec 15 12:42:05 crc kubenswrapper[4719]: I1215 12:42:05.637204 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59d59387-003d-47ae-a322-48787db59367","Type":"ContainerStarted","Data":"6b6b02db48bf17a157096ee1c0981d4e8a0236f8c08ce118e9bd1f70639f6f89"} Dec 15 12:42:05 crc kubenswrapper[4719]: I1215 12:42:05.638999 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"45c8d521-b7f2-4f93-a6e8-94d6aaf85390","Type":"ContainerStarted","Data":"0b584f5e00108a93f97cc66945d23813f46a6019e195b3df6a99392ffb4d1f5c"} Dec 15 12:42:05 crc kubenswrapper[4719]: I1215 12:42:05.663133 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.663112323 podStartE2EDuration="3.663112323s" podCreationTimestamp="2025-12-15 12:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:42:05.656405472 +0000 UTC m=+1486.598698512" watchObservedRunningTime="2025-12-15 12:42:05.663112323 +0000 UTC m=+1486.605405353" Dec 15 12:42:05 crc kubenswrapper[4719]: I1215 12:42:05.683181 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.683157333 podStartE2EDuration="3.683157333s" podCreationTimestamp="2025-12-15 12:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:42:05.679441256 +0000 UTC m=+1486.621734296" watchObservedRunningTime="2025-12-15 12:42:05.683157333 +0000 UTC m=+1486.625450383" Dec 15 12:42:07 crc kubenswrapper[4719]: I1215 12:42:07.379772 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 15 12:42:08 crc kubenswrapper[4719]: I1215 12:42:08.233202 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 15 12:42:08 crc kubenswrapper[4719]: I1215 12:42:08.233557 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 15 12:42:12 crc kubenswrapper[4719]: I1215 12:42:12.379981 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 15 12:42:12 crc kubenswrapper[4719]: I1215 12:42:12.408091 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 15 12:42:12 crc kubenswrapper[4719]: I1215 12:42:12.736398 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 15 12:42:13 crc kubenswrapper[4719]: I1215 12:42:13.233340 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 15 12:42:13 crc kubenswrapper[4719]: I1215 12:42:13.233406 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 15 12:42:13 crc kubenswrapper[4719]: I1215 12:42:13.431143 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 15 12:42:13 crc kubenswrapper[4719]: I1215 12:42:13.431884 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 15 12:42:14 crc kubenswrapper[4719]: I1215 12:42:14.247056 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="45c8d521-b7f2-4f93-a6e8-94d6aaf85390" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:42:14 crc kubenswrapper[4719]: I1215 12:42:14.247090 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="45c8d521-b7f2-4f93-a6e8-94d6aaf85390" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:42:14 crc kubenswrapper[4719]: I1215 12:42:14.443110 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="59d59387-003d-47ae-a322-48787db59367" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:42:14 crc kubenswrapper[4719]: I1215 12:42:14.443112 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="59d59387-003d-47ae-a322-48787db59367" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 15 12:42:15 crc kubenswrapper[4719]: I1215 12:42:15.857091 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 15 12:42:21 crc kubenswrapper[4719]: I1215 12:42:21.356744 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:42:21 crc kubenswrapper[4719]: I1215 12:42:21.357389 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:42:23 crc kubenswrapper[4719]: I1215 12:42:23.238153 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 15 12:42:23 crc kubenswrapper[4719]: I1215 12:42:23.239842 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 15 12:42:23 crc kubenswrapper[4719]: I1215 12:42:23.244059 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 15 12:42:23 crc kubenswrapper[4719]: I1215 12:42:23.438688 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 15 12:42:23 crc kubenswrapper[4719]: I1215 12:42:23.439379 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 15 12:42:23 crc kubenswrapper[4719]: I1215 12:42:23.439948 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 15 12:42:23 crc kubenswrapper[4719]: I1215 12:42:23.445718 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 15 12:42:23 crc kubenswrapper[4719]: I1215 12:42:23.811846 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 15 12:42:23 crc kubenswrapper[4719]: I1215 12:42:23.815126 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 15 12:42:23 crc kubenswrapper[4719]: I1215 12:42:23.822363 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 15 12:42:32 crc kubenswrapper[4719]: I1215 12:42:32.440254 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 15 12:42:33 crc kubenswrapper[4719]: I1215 12:42:33.457365 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 15 12:42:37 crc kubenswrapper[4719]: I1215 12:42:37.285529 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="62a17473-57a0-4c96-904d-a2f68bfbc92f" containerName="rabbitmq" containerID="cri-o://25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7" gracePeriod=604796 Dec 15 12:42:37 crc kubenswrapper[4719]: I1215 12:42:37.647604 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="72f88cad-2122-46b3-9493-9f5f89890956" containerName="rabbitmq" containerID="cri-o://5ffabd2ad1c6f430ef44131d3f76acd7216ae8256435b2ac1941539f45886823" gracePeriod=604796 Dec 15 12:42:38 crc kubenswrapper[4719]: I1215 12:42:38.199587 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="62a17473-57a0-4c96-904d-a2f68bfbc92f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 15 12:42:38 crc kubenswrapper[4719]: I1215 12:42:38.897480 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="72f88cad-2122-46b3-9493-9f5f89890956" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Dec 15 12:42:43 crc kubenswrapper[4719]: I1215 12:42:43.938886 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.006021 4719 generic.go:334] "Generic (PLEG): container finished" podID="72f88cad-2122-46b3-9493-9f5f89890956" containerID="5ffabd2ad1c6f430ef44131d3f76acd7216ae8256435b2ac1941539f45886823" exitCode=0 Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.006087 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"72f88cad-2122-46b3-9493-9f5f89890956","Type":"ContainerDied","Data":"5ffabd2ad1c6f430ef44131d3f76acd7216ae8256435b2ac1941539f45886823"} Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.029374 4719 generic.go:334] "Generic (PLEG): container finished" podID="62a17473-57a0-4c96-904d-a2f68bfbc92f" containerID="25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7" exitCode=0 Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.029418 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"62a17473-57a0-4c96-904d-a2f68bfbc92f","Type":"ContainerDied","Data":"25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7"} Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.029446 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"62a17473-57a0-4c96-904d-a2f68bfbc92f","Type":"ContainerDied","Data":"6515365ba4cd3df6a23b4ae7bdaa14e347b5356fbfa4177a18f91cc7bb4ffda9"} Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.029462 4719 scope.go:117] "RemoveContainer" containerID="25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.029629 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.075530 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/62a17473-57a0-4c96-904d-a2f68bfbc92f-erlang-cookie-secret\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.075693 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-erlang-cookie\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.075764 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-server-conf\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.075790 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxtmq\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-kube-api-access-fxtmq\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.075808 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-confd\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.075835 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-tls\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.075909 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-plugins-conf\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.075948 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.075995 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/62a17473-57a0-4c96-904d-a2f68bfbc92f-pod-info\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.076023 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-plugins\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.076044 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-config-data\") pod \"62a17473-57a0-4c96-904d-a2f68bfbc92f\" (UID: \"62a17473-57a0-4c96-904d-a2f68bfbc92f\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.078218 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.079493 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.079608 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.096204 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/62a17473-57a0-4c96-904d-a2f68bfbc92f-pod-info" (OuterVolumeSpecName: "pod-info") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.111629 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-kube-api-access-fxtmq" (OuterVolumeSpecName: "kube-api-access-fxtmq") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "kube-api-access-fxtmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.122093 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62a17473-57a0-4c96-904d-a2f68bfbc92f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.130603 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.157713 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.178466 4719 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.178516 4719 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.178528 4719 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/62a17473-57a0-4c96-904d-a2f68bfbc92f-pod-info\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.178540 4719 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.178553 4719 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/62a17473-57a0-4c96-904d-a2f68bfbc92f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.178566 4719 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.178581 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxtmq\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-kube-api-access-fxtmq\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.178592 4719 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.179122 4719 scope.go:117] "RemoveContainer" containerID="0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.214514 4719 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.236746 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-config-data" (OuterVolumeSpecName: "config-data") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.238463 4719 scope.go:117] "RemoveContainer" containerID="25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7" Dec 15 12:42:44 crc kubenswrapper[4719]: E1215 12:42:44.240539 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7\": container with ID starting with 25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7 not found: ID does not exist" containerID="25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.240681 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7"} err="failed to get container status \"25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7\": rpc error: code = NotFound desc = could not find container \"25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7\": container with ID starting with 25706bbdc90c5da51325d4f53c61d10a9c142a87976209907c774d0ed45777b7 not found: ID does not exist" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.240829 4719 scope.go:117] "RemoveContainer" containerID="0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7" Dec 15 12:42:44 crc kubenswrapper[4719]: E1215 12:42:44.244531 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7\": container with ID starting with 0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7 not found: ID does not exist" containerID="0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.244718 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7"} err="failed to get container status \"0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7\": rpc error: code = NotFound desc = could not find container \"0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7\": container with ID starting with 0f2fca5829cb4d783c289456e8cac600ab0cf509d43bbd6e0690d232f7a17ff7 not found: ID does not exist" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.255792 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-server-conf" (OuterVolumeSpecName: "server-conf") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.282634 4719 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.282679 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.282691 4719 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/62a17473-57a0-4c96-904d-a2f68bfbc92f-server-conf\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.293517 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.345687 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "62a17473-57a0-4c96-904d-a2f68bfbc92f" (UID: "62a17473-57a0-4c96-904d-a2f68bfbc92f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384497 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/72f88cad-2122-46b3-9493-9f5f89890956-pod-info\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384598 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-tls\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384698 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384724 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkfnj\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-kube-api-access-tkfnj\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384789 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-config-data\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384820 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-erlang-cookie\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384841 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-server-conf\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384882 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/72f88cad-2122-46b3-9493-9f5f89890956-erlang-cookie-secret\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384905 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-plugins\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384925 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-plugins-conf\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.384942 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-confd\") pod \"72f88cad-2122-46b3-9493-9f5f89890956\" (UID: \"72f88cad-2122-46b3-9493-9f5f89890956\") " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.385294 4719 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/62a17473-57a0-4c96-904d-a2f68bfbc92f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.388633 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.392761 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.392914 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.392996 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.399339 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f88cad-2122-46b3-9493-9f5f89890956-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.399661 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-kube-api-access-tkfnj" (OuterVolumeSpecName: "kube-api-access-tkfnj") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "kube-api-access-tkfnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.402117 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.406458 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/72f88cad-2122-46b3-9493-9f5f89890956-pod-info" (OuterVolumeSpecName: "pod-info") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.469807 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-config-data" (OuterVolumeSpecName: "config-data") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.489831 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-server-conf" (OuterVolumeSpecName: "server-conf") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.490192 4719 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-server-conf\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.490225 4719 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/72f88cad-2122-46b3-9493-9f5f89890956-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.490237 4719 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.490246 4719 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.490293 4719 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/72f88cad-2122-46b3-9493-9f5f89890956-pod-info\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.490302 4719 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.490327 4719 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.490347 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkfnj\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-kube-api-access-tkfnj\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.490356 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/72f88cad-2122-46b3-9493-9f5f89890956-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.490367 4719 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.591517 4719 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.592085 4719 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.662796 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "72f88cad-2122-46b3-9493-9f5f89890956" (UID: "72f88cad-2122-46b3-9493-9f5f89890956"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.695309 4719 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/72f88cad-2122-46b3-9493-9f5f89890956-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.744761 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6cgf5"] Dec 15 12:42:44 crc kubenswrapper[4719]: E1215 12:42:44.745144 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62a17473-57a0-4c96-904d-a2f68bfbc92f" containerName="rabbitmq" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.745160 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="62a17473-57a0-4c96-904d-a2f68bfbc92f" containerName="rabbitmq" Dec 15 12:42:44 crc kubenswrapper[4719]: E1215 12:42:44.745182 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62a17473-57a0-4c96-904d-a2f68bfbc92f" containerName="setup-container" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.745189 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="62a17473-57a0-4c96-904d-a2f68bfbc92f" containerName="setup-container" Dec 15 12:42:44 crc kubenswrapper[4719]: E1215 12:42:44.745200 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f88cad-2122-46b3-9493-9f5f89890956" containerName="rabbitmq" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.745207 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f88cad-2122-46b3-9493-9f5f89890956" containerName="rabbitmq" Dec 15 12:42:44 crc kubenswrapper[4719]: E1215 12:42:44.745221 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f88cad-2122-46b3-9493-9f5f89890956" containerName="setup-container" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.745227 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f88cad-2122-46b3-9493-9f5f89890956" containerName="setup-container" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.745376 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="62a17473-57a0-4c96-904d-a2f68bfbc92f" containerName="rabbitmq" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.745387 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="72f88cad-2122-46b3-9493-9f5f89890956" containerName="rabbitmq" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.751131 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.757285 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.757880 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6cgf5"] Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.765833 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.833616 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.886541 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.888345 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.892496 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.892638 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.892801 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.892883 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.892516 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.893455 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mwhq7" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.893632 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.899907 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.916958 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-config\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.917054 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.917120 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.917155 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.917184 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkf2g\" (UniqueName: \"kubernetes.io/projected/0e747d7d-07ff-49a8-9101-97cd1635908c-kube-api-access-jkf2g\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.917212 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-svc\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.917228 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.954643 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6cgf5"] Dec 15 12:42:44 crc kubenswrapper[4719]: E1215 12:42:44.955351 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-jkf2g openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" podUID="0e747d7d-07ff-49a8-9101-97cd1635908c" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.978959 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-ll29b"] Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.980332 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:44 crc kubenswrapper[4719]: I1215 12:42:44.996417 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-ll29b"] Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019389 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d4163ed6-9550-4d10-baab-b92859fa07df-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019487 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019512 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d4163ed6-9550-4d10-baab-b92859fa07df-config-data\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019545 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019583 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019613 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkf2g\" (UniqueName: \"kubernetes.io/projected/0e747d7d-07ff-49a8-9101-97cd1635908c-kube-api-access-jkf2g\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019647 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d4163ed6-9550-4d10-baab-b92859fa07df-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019674 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019696 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-svc\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019730 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019756 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-config\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019803 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.019937 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.020693 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-svc\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.020925 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.020979 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.021205 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d4163ed6-9550-4d10-baab-b92859fa07df-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.021287 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds5dn\" (UniqueName: \"kubernetes.io/projected/d4163ed6-9550-4d10-baab-b92859fa07df-kube-api-access-ds5dn\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.021316 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-config\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.021410 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.021427 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.021560 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.021589 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d4163ed6-9550-4d10-baab-b92859fa07df-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.022353 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.038796 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"72f88cad-2122-46b3-9493-9f5f89890956","Type":"ContainerDied","Data":"f1b3d66ec78df67cd57ae05ead32f7f4ac543fb0ead7f63c4ec0d91873a25b46"} Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.038935 4719 scope.go:117] "RemoveContainer" containerID="5ffabd2ad1c6f430ef44131d3f76acd7216ae8256435b2ac1941539f45886823" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.039115 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.045376 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.049695 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkf2g\" (UniqueName: \"kubernetes.io/projected/0e747d7d-07ff-49a8-9101-97cd1635908c-kube-api-access-jkf2g\") pod \"dnsmasq-dns-67b789f86c-6cgf5\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.095147 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.100304 4719 scope.go:117] "RemoveContainer" containerID="6e34719f6e20b66097cb87d45428b9ef0d5203fb97ac81e566f14c57bc6c4811" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.117449 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123124 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d4163ed6-9550-4d10-baab-b92859fa07df-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123206 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-config\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123252 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d4163ed6-9550-4d10-baab-b92859fa07df-config-data\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123297 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123337 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123371 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123401 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123427 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d4163ed6-9550-4d10-baab-b92859fa07df-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123458 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123752 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123816 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpj5m\" (UniqueName: \"kubernetes.io/projected/3b63219e-0ba3-49f4-9a94-72a61b2660f3-kube-api-access-rpj5m\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.123845 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.124041 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.124070 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d4163ed6-9550-4d10-baab-b92859fa07df-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.124096 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds5dn\" (UniqueName: \"kubernetes.io/projected/d4163ed6-9550-4d10-baab-b92859fa07df-kube-api-access-ds5dn\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.124127 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.124180 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.124203 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d4163ed6-9550-4d10-baab-b92859fa07df-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.124250 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d4163ed6-9550-4d10-baab-b92859fa07df-config-data\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.124296 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.124647 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.124937 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d4163ed6-9550-4d10-baab-b92859fa07df-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.125185 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d4163ed6-9550-4d10-baab-b92859fa07df-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.125205 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.132808 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d4163ed6-9550-4d10-baab-b92859fa07df-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.133979 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d4163ed6-9550-4d10-baab-b92859fa07df-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.134838 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.144738 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.151003 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d4163ed6-9550-4d10-baab-b92859fa07df-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.154843 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.157821 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds5dn\" (UniqueName: \"kubernetes.io/projected/d4163ed6-9550-4d10-baab-b92859fa07df-kube-api-access-ds5dn\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.164101 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.169530 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.169823 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.170047 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.170230 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.170333 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.170435 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2df6g" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.170942 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.174688 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.193560 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"d4163ed6-9550-4d10-baab-b92859fa07df\") " pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.215474 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.225178 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-svc\") pod \"0e747d7d-07ff-49a8-9101-97cd1635908c\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.225967 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkf2g\" (UniqueName: \"kubernetes.io/projected/0e747d7d-07ff-49a8-9101-97cd1635908c-kube-api-access-jkf2g\") pod \"0e747d7d-07ff-49a8-9101-97cd1635908c\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.226111 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-config\") pod \"0e747d7d-07ff-49a8-9101-97cd1635908c\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.226220 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-nb\") pod \"0e747d7d-07ff-49a8-9101-97cd1635908c\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.226309 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-swift-storage-0\") pod \"0e747d7d-07ff-49a8-9101-97cd1635908c\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.226911 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-openstack-edpm-ipam\") pod \"0e747d7d-07ff-49a8-9101-97cd1635908c\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.227035 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-sb\") pod \"0e747d7d-07ff-49a8-9101-97cd1635908c\" (UID: \"0e747d7d-07ff-49a8-9101-97cd1635908c\") " Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.225895 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0e747d7d-07ff-49a8-9101-97cd1635908c" (UID: "0e747d7d-07ff-49a8-9101-97cd1635908c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.226548 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-config" (OuterVolumeSpecName: "config") pod "0e747d7d-07ff-49a8-9101-97cd1635908c" (UID: "0e747d7d-07ff-49a8-9101-97cd1635908c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.226721 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0e747d7d-07ff-49a8-9101-97cd1635908c" (UID: "0e747d7d-07ff-49a8-9101-97cd1635908c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.226913 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0e747d7d-07ff-49a8-9101-97cd1635908c" (UID: "0e747d7d-07ff-49a8-9101-97cd1635908c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.227183 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "0e747d7d-07ff-49a8-9101-97cd1635908c" (UID: "0e747d7d-07ff-49a8-9101-97cd1635908c"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.227541 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-config\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.227715 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.227805 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228165 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228266 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228377 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpj5m\" (UniqueName: \"kubernetes.io/projected/3b63219e-0ba3-49f4-9a94-72a61b2660f3-kube-api-access-rpj5m\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228497 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-config\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228504 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228744 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228760 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228771 4719 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228781 4719 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228790 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.227589 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0e747d7d-07ff-49a8-9101-97cd1635908c" (UID: "0e747d7d-07ff-49a8-9101-97cd1635908c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.228933 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e747d7d-07ff-49a8-9101-97cd1635908c-kube-api-access-jkf2g" (OuterVolumeSpecName: "kube-api-access-jkf2g") pod "0e747d7d-07ff-49a8-9101-97cd1635908c" (UID: "0e747d7d-07ff-49a8-9101-97cd1635908c"). InnerVolumeSpecName "kube-api-access-jkf2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.229439 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.229638 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.229713 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.230035 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.230372 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.248253 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpj5m\" (UniqueName: \"kubernetes.io/projected/3b63219e-0ba3-49f4-9a94-72a61b2660f3-kube-api-access-rpj5m\") pod \"dnsmasq-dns-cb6ffcf87-ll29b\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.295783 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.333382 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334174 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82c8c133-e1d6-4b56-b745-1aed017d0bfb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334225 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334279 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t4t7\" (UniqueName: \"kubernetes.io/projected/82c8c133-e1d6-4b56-b745-1aed017d0bfb-kube-api-access-7t4t7\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334342 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82c8c133-e1d6-4b56-b745-1aed017d0bfb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334408 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82c8c133-e1d6-4b56-b745-1aed017d0bfb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334434 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82c8c133-e1d6-4b56-b745-1aed017d0bfb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334479 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334505 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82c8c133-e1d6-4b56-b745-1aed017d0bfb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334615 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334655 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334765 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkf2g\" (UniqueName: \"kubernetes.io/projected/0e747d7d-07ff-49a8-9101-97cd1635908c-kube-api-access-jkf2g\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.334780 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e747d7d-07ff-49a8-9101-97cd1635908c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.436812 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t4t7\" (UniqueName: \"kubernetes.io/projected/82c8c133-e1d6-4b56-b745-1aed017d0bfb-kube-api-access-7t4t7\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.436877 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82c8c133-e1d6-4b56-b745-1aed017d0bfb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.436913 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82c8c133-e1d6-4b56-b745-1aed017d0bfb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.436932 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82c8c133-e1d6-4b56-b745-1aed017d0bfb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.436955 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.436972 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82c8c133-e1d6-4b56-b745-1aed017d0bfb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.437049 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.437110 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.437164 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.437197 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82c8c133-e1d6-4b56-b745-1aed017d0bfb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.437220 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.437772 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.439926 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.441267 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82c8c133-e1d6-4b56-b745-1aed017d0bfb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.441669 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82c8c133-e1d6-4b56-b745-1aed017d0bfb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.443125 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.443483 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82c8c133-e1d6-4b56-b745-1aed017d0bfb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.462653 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82c8c133-e1d6-4b56-b745-1aed017d0bfb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.474591 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.481496 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82c8c133-e1d6-4b56-b745-1aed017d0bfb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.481998 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82c8c133-e1d6-4b56-b745-1aed017d0bfb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.485535 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t4t7\" (UniqueName: \"kubernetes.io/projected/82c8c133-e1d6-4b56-b745-1aed017d0bfb-kube-api-access-7t4t7\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.535904 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62a17473-57a0-4c96-904d-a2f68bfbc92f" path="/var/lib/kubelet/pods/62a17473-57a0-4c96-904d-a2f68bfbc92f/volumes" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.537475 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72f88cad-2122-46b3-9493-9f5f89890956" path="/var/lib/kubelet/pods/72f88cad-2122-46b3-9493-9f5f89890956/volumes" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.628126 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"82c8c133-e1d6-4b56-b745-1aed017d0bfb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.673399 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 15 12:42:45 crc kubenswrapper[4719]: I1215 12:42:45.792078 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:42:46 crc kubenswrapper[4719]: I1215 12:42:46.057905 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d4163ed6-9550-4d10-baab-b92859fa07df","Type":"ContainerStarted","Data":"ad263016915d603ee9231394237503c3b99645fb9d44f0451ca5afbd982f8e8e"} Dec 15 12:42:46 crc kubenswrapper[4719]: I1215 12:42:46.058047 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-6cgf5" Dec 15 12:42:46 crc kubenswrapper[4719]: I1215 12:42:46.116772 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6cgf5"] Dec 15 12:42:46 crc kubenswrapper[4719]: I1215 12:42:46.133661 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-6cgf5"] Dec 15 12:42:46 crc kubenswrapper[4719]: I1215 12:42:46.143150 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-ll29b"] Dec 15 12:42:46 crc kubenswrapper[4719]: W1215 12:42:46.399135 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82c8c133_e1d6_4b56_b745_1aed017d0bfb.slice/crio-8cfe7ada0c2b65aae80d0107e6d6f79c1ffd3efaacef498e7c738cdee459e1b8 WatchSource:0}: Error finding container 8cfe7ada0c2b65aae80d0107e6d6f79c1ffd3efaacef498e7c738cdee459e1b8: Status 404 returned error can't find the container with id 8cfe7ada0c2b65aae80d0107e6d6f79c1ffd3efaacef498e7c738cdee459e1b8 Dec 15 12:42:46 crc kubenswrapper[4719]: I1215 12:42:46.409095 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 15 12:42:47 crc kubenswrapper[4719]: I1215 12:42:47.067550 4719 generic.go:334] "Generic (PLEG): container finished" podID="3b63219e-0ba3-49f4-9a94-72a61b2660f3" containerID="183704a30dd55912f22c2aab39a9718bc9f7d43837cf3f162f95f8754c6e3b56" exitCode=0 Dec 15 12:42:47 crc kubenswrapper[4719]: I1215 12:42:47.067650 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" event={"ID":"3b63219e-0ba3-49f4-9a94-72a61b2660f3","Type":"ContainerDied","Data":"183704a30dd55912f22c2aab39a9718bc9f7d43837cf3f162f95f8754c6e3b56"} Dec 15 12:42:47 crc kubenswrapper[4719]: I1215 12:42:47.067917 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" event={"ID":"3b63219e-0ba3-49f4-9a94-72a61b2660f3","Type":"ContainerStarted","Data":"a70d18a614828f47e958fe3d71d39e18f1a702af8451a4601791d56b3a673e5f"} Dec 15 12:42:47 crc kubenswrapper[4719]: I1215 12:42:47.069103 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82c8c133-e1d6-4b56-b745-1aed017d0bfb","Type":"ContainerStarted","Data":"8cfe7ada0c2b65aae80d0107e6d6f79c1ffd3efaacef498e7c738cdee459e1b8"} Dec 15 12:42:47 crc kubenswrapper[4719]: I1215 12:42:47.490712 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e747d7d-07ff-49a8-9101-97cd1635908c" path="/var/lib/kubelet/pods/0e747d7d-07ff-49a8-9101-97cd1635908c/volumes" Dec 15 12:42:48 crc kubenswrapper[4719]: I1215 12:42:48.081818 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d4163ed6-9550-4d10-baab-b92859fa07df","Type":"ContainerStarted","Data":"fc6b0ea4a4c59618d24635bbd4f1ef8f657ddcd7f59d9cf42897d7cecb1b05e0"} Dec 15 12:42:48 crc kubenswrapper[4719]: I1215 12:42:48.086080 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" event={"ID":"3b63219e-0ba3-49f4-9a94-72a61b2660f3","Type":"ContainerStarted","Data":"6eb0c09f89f7994e5fba0976d4cb2559058f8adbac3e95b2535d20f16280dfa4"} Dec 15 12:42:48 crc kubenswrapper[4719]: I1215 12:42:48.086392 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:48 crc kubenswrapper[4719]: I1215 12:42:48.088168 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82c8c133-e1d6-4b56-b745-1aed017d0bfb","Type":"ContainerStarted","Data":"482154fdb3ab3a864de58c26e381cce6ac81d35337b60e1d8c3b0453b6f9b90a"} Dec 15 12:42:48 crc kubenswrapper[4719]: I1215 12:42:48.174144 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" podStartSLOduration=4.174118488 podStartE2EDuration="4.174118488s" podCreationTimestamp="2025-12-15 12:42:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:42:48.168315075 +0000 UTC m=+1529.110608145" watchObservedRunningTime="2025-12-15 12:42:48.174118488 +0000 UTC m=+1529.116411528" Dec 15 12:42:51 crc kubenswrapper[4719]: I1215 12:42:51.357317 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:42:51 crc kubenswrapper[4719]: I1215 12:42:51.357743 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.298053 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.365040 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-psnpv"] Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.365301 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" podUID="2fd5121b-dd23-4935-a6d2-3a81e92193cf" containerName="dnsmasq-dns" containerID="cri-o://cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685" gracePeriod=10 Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.591487 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-cvlz2"] Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.593263 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.618763 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-cvlz2"] Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.672991 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.673049 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjfvh\" (UniqueName: \"kubernetes.io/projected/3cd29959-bbda-441f-901a-726fc4ebb4c4-kube-api-access-wjfvh\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.673073 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.673101 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.673136 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-config\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.673158 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.673185 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.774439 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-config\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.775433 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.775482 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.775585 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.775621 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjfvh\" (UniqueName: \"kubernetes.io/projected/3cd29959-bbda-441f-901a-726fc4ebb4c4-kube-api-access-wjfvh\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.775648 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.775678 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.776488 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.776804 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.777016 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.777347 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-config\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.777792 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.777895 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3cd29959-bbda-441f-901a-726fc4ebb4c4-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.799060 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjfvh\" (UniqueName: \"kubernetes.io/projected/3cd29959-bbda-441f-901a-726fc4ebb4c4-kube-api-access-wjfvh\") pod \"dnsmasq-dns-79dc84bdb7-cvlz2\" (UID: \"3cd29959-bbda-441f-901a-726fc4ebb4c4\") " pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:55 crc kubenswrapper[4719]: I1215 12:42:55.919055 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.004566 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.083722 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-nb\") pod \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.084163 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-sb\") pod \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.084349 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-svc\") pod \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.084379 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdcws\" (UniqueName: \"kubernetes.io/projected/2fd5121b-dd23-4935-a6d2-3a81e92193cf-kube-api-access-jdcws\") pod \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.084417 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-swift-storage-0\") pod \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.084570 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-config\") pod \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\" (UID: \"2fd5121b-dd23-4935-a6d2-3a81e92193cf\") " Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.100874 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fd5121b-dd23-4935-a6d2-3a81e92193cf-kube-api-access-jdcws" (OuterVolumeSpecName: "kube-api-access-jdcws") pod "2fd5121b-dd23-4935-a6d2-3a81e92193cf" (UID: "2fd5121b-dd23-4935-a6d2-3a81e92193cf"). InnerVolumeSpecName "kube-api-access-jdcws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.142203 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-config" (OuterVolumeSpecName: "config") pod "2fd5121b-dd23-4935-a6d2-3a81e92193cf" (UID: "2fd5121b-dd23-4935-a6d2-3a81e92193cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.142298 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2fd5121b-dd23-4935-a6d2-3a81e92193cf" (UID: "2fd5121b-dd23-4935-a6d2-3a81e92193cf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.146436 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2fd5121b-dd23-4935-a6d2-3a81e92193cf" (UID: "2fd5121b-dd23-4935-a6d2-3a81e92193cf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.158425 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2fd5121b-dd23-4935-a6d2-3a81e92193cf" (UID: "2fd5121b-dd23-4935-a6d2-3a81e92193cf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.164611 4719 generic.go:334] "Generic (PLEG): container finished" podID="2fd5121b-dd23-4935-a6d2-3a81e92193cf" containerID="cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685" exitCode=0 Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.164676 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.164697 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" event={"ID":"2fd5121b-dd23-4935-a6d2-3a81e92193cf","Type":"ContainerDied","Data":"cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685"} Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.166148 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-psnpv" event={"ID":"2fd5121b-dd23-4935-a6d2-3a81e92193cf","Type":"ContainerDied","Data":"92088af17ad5bcfff853a8824caab0b37d1a3265a349883a837d4a7c262a0050"} Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.166199 4719 scope.go:117] "RemoveContainer" containerID="cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.168479 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2fd5121b-dd23-4935-a6d2-3a81e92193cf" (UID: "2fd5121b-dd23-4935-a6d2-3a81e92193cf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.189328 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.189365 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.189380 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.189392 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.189408 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdcws\" (UniqueName: \"kubernetes.io/projected/2fd5121b-dd23-4935-a6d2-3a81e92193cf-kube-api-access-jdcws\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.189420 4719 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fd5121b-dd23-4935-a6d2-3a81e92193cf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.196536 4719 scope.go:117] "RemoveContainer" containerID="077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.223492 4719 scope.go:117] "RemoveContainer" containerID="cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685" Dec 15 12:42:56 crc kubenswrapper[4719]: E1215 12:42:56.224168 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685\": container with ID starting with cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685 not found: ID does not exist" containerID="cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.224233 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685"} err="failed to get container status \"cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685\": rpc error: code = NotFound desc = could not find container \"cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685\": container with ID starting with cb5fb227f11a91956f811779f71936066216f1a299ff0c9af49dba30c4ba9685 not found: ID does not exist" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.224261 4719 scope.go:117] "RemoveContainer" containerID="077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a" Dec 15 12:42:56 crc kubenswrapper[4719]: E1215 12:42:56.224597 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a\": container with ID starting with 077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a not found: ID does not exist" containerID="077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.224642 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a"} err="failed to get container status \"077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a\": rpc error: code = NotFound desc = could not find container \"077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a\": container with ID starting with 077925c3380182661d8dbd15ee3a8a715647ff5dab0f31a7b661252d7bc6806a not found: ID does not exist" Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.425246 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-cvlz2"] Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.618188 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-psnpv"] Dec 15 12:42:56 crc kubenswrapper[4719]: I1215 12:42:56.625815 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-psnpv"] Dec 15 12:42:57 crc kubenswrapper[4719]: I1215 12:42:57.181758 4719 generic.go:334] "Generic (PLEG): container finished" podID="3cd29959-bbda-441f-901a-726fc4ebb4c4" containerID="473eb64dfd79368403709524ed825174860faab8a07bd3c8bbbac98975270ad6" exitCode=0 Dec 15 12:42:57 crc kubenswrapper[4719]: I1215 12:42:57.181848 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" event={"ID":"3cd29959-bbda-441f-901a-726fc4ebb4c4","Type":"ContainerDied","Data":"473eb64dfd79368403709524ed825174860faab8a07bd3c8bbbac98975270ad6"} Dec 15 12:42:57 crc kubenswrapper[4719]: I1215 12:42:57.182210 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" event={"ID":"3cd29959-bbda-441f-901a-726fc4ebb4c4","Type":"ContainerStarted","Data":"3565b5fb93d0dbb5a083a8ac444f6e264f2251e9a707b988074965075669bcc9"} Dec 15 12:42:57 crc kubenswrapper[4719]: I1215 12:42:57.486695 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fd5121b-dd23-4935-a6d2-3a81e92193cf" path="/var/lib/kubelet/pods/2fd5121b-dd23-4935-a6d2-3a81e92193cf/volumes" Dec 15 12:42:58 crc kubenswrapper[4719]: I1215 12:42:58.195202 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" event={"ID":"3cd29959-bbda-441f-901a-726fc4ebb4c4","Type":"ContainerStarted","Data":"c844e463371b32052f8c6b45ece074742c5eafd902f5d11b52be34fba23fadce"} Dec 15 12:42:58 crc kubenswrapper[4719]: I1215 12:42:58.195371 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:42:58 crc kubenswrapper[4719]: I1215 12:42:58.224727 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" podStartSLOduration=3.224709918 podStartE2EDuration="3.224709918s" podCreationTimestamp="2025-12-15 12:42:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:42:58.212556776 +0000 UTC m=+1539.154849806" watchObservedRunningTime="2025-12-15 12:42:58.224709918 +0000 UTC m=+1539.167002948" Dec 15 12:43:05 crc kubenswrapper[4719]: I1215 12:43:05.920992 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79dc84bdb7-cvlz2" Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.001066 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-ll29b"] Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.001368 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" podUID="3b63219e-0ba3-49f4-9a94-72a61b2660f3" containerName="dnsmasq-dns" containerID="cri-o://6eb0c09f89f7994e5fba0976d4cb2559058f8adbac3e95b2535d20f16280dfa4" gracePeriod=10 Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.347722 4719 generic.go:334] "Generic (PLEG): container finished" podID="3b63219e-0ba3-49f4-9a94-72a61b2660f3" containerID="6eb0c09f89f7994e5fba0976d4cb2559058f8adbac3e95b2535d20f16280dfa4" exitCode=0 Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.347753 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" event={"ID":"3b63219e-0ba3-49f4-9a94-72a61b2660f3","Type":"ContainerDied","Data":"6eb0c09f89f7994e5fba0976d4cb2559058f8adbac3e95b2535d20f16280dfa4"} Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.778013 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.927258 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpj5m\" (UniqueName: \"kubernetes.io/projected/3b63219e-0ba3-49f4-9a94-72a61b2660f3-kube-api-access-rpj5m\") pod \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.928142 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-swift-storage-0\") pod \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.928487 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-sb\") pod \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.928573 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-nb\") pod \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.928919 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-svc\") pod \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.928987 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-openstack-edpm-ipam\") pod \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.929176 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-config\") pod \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\" (UID: \"3b63219e-0ba3-49f4-9a94-72a61b2660f3\") " Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.968873 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b63219e-0ba3-49f4-9a94-72a61b2660f3-kube-api-access-rpj5m" (OuterVolumeSpecName: "kube-api-access-rpj5m") pod "3b63219e-0ba3-49f4-9a94-72a61b2660f3" (UID: "3b63219e-0ba3-49f4-9a94-72a61b2660f3"). InnerVolumeSpecName "kube-api-access-rpj5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:43:06 crc kubenswrapper[4719]: I1215 12:43:06.987207 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3b63219e-0ba3-49f4-9a94-72a61b2660f3" (UID: "3b63219e-0ba3-49f4-9a94-72a61b2660f3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.001476 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3b63219e-0ba3-49f4-9a94-72a61b2660f3" (UID: "3b63219e-0ba3-49f4-9a94-72a61b2660f3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.018494 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3b63219e-0ba3-49f4-9a94-72a61b2660f3" (UID: "3b63219e-0ba3-49f4-9a94-72a61b2660f3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.019225 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "3b63219e-0ba3-49f4-9a94-72a61b2660f3" (UID: "3b63219e-0ba3-49f4-9a94-72a61b2660f3"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.035409 4719 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.035436 4719 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.035448 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpj5m\" (UniqueName: \"kubernetes.io/projected/3b63219e-0ba3-49f4-9a94-72a61b2660f3-kube-api-access-rpj5m\") on node \"crc\" DevicePath \"\"" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.035459 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.035471 4719 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.037604 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-config" (OuterVolumeSpecName: "config") pod "3b63219e-0ba3-49f4-9a94-72a61b2660f3" (UID: "3b63219e-0ba3-49f4-9a94-72a61b2660f3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.043308 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3b63219e-0ba3-49f4-9a94-72a61b2660f3" (UID: "3b63219e-0ba3-49f4-9a94-72a61b2660f3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.136890 4719 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-config\") on node \"crc\" DevicePath \"\"" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.136917 4719 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b63219e-0ba3-49f4-9a94-72a61b2660f3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.358168 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" event={"ID":"3b63219e-0ba3-49f4-9a94-72a61b2660f3","Type":"ContainerDied","Data":"a70d18a614828f47e958fe3d71d39e18f1a702af8451a4601791d56b3a673e5f"} Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.358222 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-ll29b" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.358242 4719 scope.go:117] "RemoveContainer" containerID="6eb0c09f89f7994e5fba0976d4cb2559058f8adbac3e95b2535d20f16280dfa4" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.394105 4719 scope.go:117] "RemoveContainer" containerID="183704a30dd55912f22c2aab39a9718bc9f7d43837cf3f162f95f8754c6e3b56" Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.400219 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-ll29b"] Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.420684 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-ll29b"] Dec 15 12:43:07 crc kubenswrapper[4719]: I1215 12:43:07.486440 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b63219e-0ba3-49f4-9a94-72a61b2660f3" path="/var/lib/kubelet/pods/3b63219e-0ba3-49f4-9a94-72a61b2660f3/volumes" Dec 15 12:43:20 crc kubenswrapper[4719]: I1215 12:43:20.503078 4719 generic.go:334] "Generic (PLEG): container finished" podID="82c8c133-e1d6-4b56-b745-1aed017d0bfb" containerID="482154fdb3ab3a864de58c26e381cce6ac81d35337b60e1d8c3b0453b6f9b90a" exitCode=0 Dec 15 12:43:20 crc kubenswrapper[4719]: I1215 12:43:20.503142 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82c8c133-e1d6-4b56-b745-1aed017d0bfb","Type":"ContainerDied","Data":"482154fdb3ab3a864de58c26e381cce6ac81d35337b60e1d8c3b0453b6f9b90a"} Dec 15 12:43:20 crc kubenswrapper[4719]: I1215 12:43:20.507741 4719 generic.go:334] "Generic (PLEG): container finished" podID="d4163ed6-9550-4d10-baab-b92859fa07df" containerID="fc6b0ea4a4c59618d24635bbd4f1ef8f657ddcd7f59d9cf42897d7cecb1b05e0" exitCode=0 Dec 15 12:43:20 crc kubenswrapper[4719]: I1215 12:43:20.507782 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d4163ed6-9550-4d10-baab-b92859fa07df","Type":"ContainerDied","Data":"fc6b0ea4a4c59618d24635bbd4f1ef8f657ddcd7f59d9cf42897d7cecb1b05e0"} Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.357293 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.357608 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.357650 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.358332 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.358383 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" gracePeriod=600 Dec 15 12:43:21 crc kubenswrapper[4719]: E1215 12:43:21.541554 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.570249 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"82c8c133-e1d6-4b56-b745-1aed017d0bfb","Type":"ContainerStarted","Data":"e9e019d2025f0e3fb823072c44116d13a623318181012cd9e4c63baf289baf7b"} Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.571348 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.577376 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" exitCode=0 Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.577435 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b"} Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.577467 4719 scope.go:117] "RemoveContainer" containerID="e82d238054920207f8e0cb0360298c4a0049bf10cc2919ab754c599ae7031178" Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.578083 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:43:21 crc kubenswrapper[4719]: E1215 12:43:21.578301 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.592383 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d4163ed6-9550-4d10-baab-b92859fa07df","Type":"ContainerStarted","Data":"fd51fad007c7f771db9a35ae730ad94c836d30d6f56c7515b764df91e1aeb1a1"} Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.593164 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.612058 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.612036869 podStartE2EDuration="36.612036869s" podCreationTimestamp="2025-12-15 12:42:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:43:21.605980759 +0000 UTC m=+1562.548273799" watchObservedRunningTime="2025-12-15 12:43:21.612036869 +0000 UTC m=+1562.554329899" Dec 15 12:43:21 crc kubenswrapper[4719]: I1215 12:43:21.659165 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.659145798 podStartE2EDuration="37.659145798s" podCreationTimestamp="2025-12-15 12:42:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 12:43:21.653186 +0000 UTC m=+1562.595479030" watchObservedRunningTime="2025-12-15 12:43:21.659145798 +0000 UTC m=+1562.601438828" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.598595 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s"] Dec 15 12:43:29 crc kubenswrapper[4719]: E1215 12:43:29.599506 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b63219e-0ba3-49f4-9a94-72a61b2660f3" containerName="dnsmasq-dns" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.599520 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b63219e-0ba3-49f4-9a94-72a61b2660f3" containerName="dnsmasq-dns" Dec 15 12:43:29 crc kubenswrapper[4719]: E1215 12:43:29.599536 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd5121b-dd23-4935-a6d2-3a81e92193cf" containerName="init" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.599542 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd5121b-dd23-4935-a6d2-3a81e92193cf" containerName="init" Dec 15 12:43:29 crc kubenswrapper[4719]: E1215 12:43:29.599569 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b63219e-0ba3-49f4-9a94-72a61b2660f3" containerName="init" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.599575 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b63219e-0ba3-49f4-9a94-72a61b2660f3" containerName="init" Dec 15 12:43:29 crc kubenswrapper[4719]: E1215 12:43:29.599595 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd5121b-dd23-4935-a6d2-3a81e92193cf" containerName="dnsmasq-dns" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.599602 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd5121b-dd23-4935-a6d2-3a81e92193cf" containerName="dnsmasq-dns" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.599784 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b63219e-0ba3-49f4-9a94-72a61b2660f3" containerName="dnsmasq-dns" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.599810 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fd5121b-dd23-4935-a6d2-3a81e92193cf" containerName="dnsmasq-dns" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.600440 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.602848 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.603644 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.603832 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.604014 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.612615 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s"] Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.704271 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.704436 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzsrl\" (UniqueName: \"kubernetes.io/projected/588db7b9-6a66-4c02-be83-8766e47c211c-kube-api-access-pzsrl\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.704491 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.704570 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.806438 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.806566 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.806678 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzsrl\" (UniqueName: \"kubernetes.io/projected/588db7b9-6a66-4c02-be83-8766e47c211c-kube-api-access-pzsrl\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.806721 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.812125 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.815569 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.826719 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.850514 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzsrl\" (UniqueName: \"kubernetes.io/projected/588db7b9-6a66-4c02-be83-8766e47c211c-kube-api-access-pzsrl\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:29 crc kubenswrapper[4719]: I1215 12:43:29.918262 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:43:30 crc kubenswrapper[4719]: I1215 12:43:30.714573 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s"] Dec 15 12:43:31 crc kubenswrapper[4719]: I1215 12:43:31.686731 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" event={"ID":"588db7b9-6a66-4c02-be83-8766e47c211c","Type":"ContainerStarted","Data":"358f83f9666dedc4a353f2562aefe6588190c939170afcdb7a1eddc1ff6cbaf6"} Dec 15 12:43:32 crc kubenswrapper[4719]: I1215 12:43:32.193685 4719 scope.go:117] "RemoveContainer" containerID="e76c3a5c2744508c3ceaf62081538d3fff95fc0321650867c6db547d5b5e7f7d" Dec 15 12:43:32 crc kubenswrapper[4719]: I1215 12:43:32.223412 4719 scope.go:117] "RemoveContainer" containerID="afeddce05a6681af855875ccab7c2a377353193f66c8804e0f1ec4be1dd21b07" Dec 15 12:43:32 crc kubenswrapper[4719]: I1215 12:43:32.476711 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:43:32 crc kubenswrapper[4719]: E1215 12:43:32.477398 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:43:35 crc kubenswrapper[4719]: I1215 12:43:35.220064 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 15 12:43:35 crc kubenswrapper[4719]: I1215 12:43:35.795235 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 15 12:43:45 crc kubenswrapper[4719]: I1215 12:43:45.476999 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:43:45 crc kubenswrapper[4719]: E1215 12:43:45.478325 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:43:51 crc kubenswrapper[4719]: E1215 12:43:51.997808 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Dec 15 12:43:51 crc kubenswrapper[4719]: E1215 12:43:51.998784 4719 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 15 12:43:51 crc kubenswrapper[4719]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Dec 15 12:43:51 crc kubenswrapper[4719]: - hosts: all Dec 15 12:43:51 crc kubenswrapper[4719]: strategy: linear Dec 15 12:43:51 crc kubenswrapper[4719]: tasks: Dec 15 12:43:51 crc kubenswrapper[4719]: - name: Enable podified-repos Dec 15 12:43:51 crc kubenswrapper[4719]: become: true Dec 15 12:43:51 crc kubenswrapper[4719]: ansible.builtin.shell: | Dec 15 12:43:51 crc kubenswrapper[4719]: set -euxo pipefail Dec 15 12:43:51 crc kubenswrapper[4719]: pushd /var/tmp Dec 15 12:43:51 crc kubenswrapper[4719]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Dec 15 12:43:51 crc kubenswrapper[4719]: pushd repo-setup-main Dec 15 12:43:51 crc kubenswrapper[4719]: python3 -m venv ./venv Dec 15 12:43:51 crc kubenswrapper[4719]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Dec 15 12:43:51 crc kubenswrapper[4719]: ./venv/bin/repo-setup current-podified -b antelope Dec 15 12:43:51 crc kubenswrapper[4719]: popd Dec 15 12:43:51 crc kubenswrapper[4719]: rm -rf repo-setup-main Dec 15 12:43:51 crc kubenswrapper[4719]: Dec 15 12:43:51 crc kubenswrapper[4719]: Dec 15 12:43:51 crc kubenswrapper[4719]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Dec 15 12:43:51 crc kubenswrapper[4719]: edpm_override_hosts: openstack-edpm-ipam Dec 15 12:43:51 crc kubenswrapper[4719]: edpm_service_type: repo-setup Dec 15 12:43:51 crc kubenswrapper[4719]: Dec 15 12:43:51 crc kubenswrapper[4719]: Dec 15 12:43:51 crc kubenswrapper[4719]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pzsrl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s_openstack(588db7b9-6a66-4c02-be83-8766e47c211c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Dec 15 12:43:51 crc kubenswrapper[4719]: > logger="UnhandledError" Dec 15 12:43:52 crc kubenswrapper[4719]: E1215 12:43:52.000072 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" podUID="588db7b9-6a66-4c02-be83-8766e47c211c" Dec 15 12:43:52 crc kubenswrapper[4719]: E1215 12:43:52.909747 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" podUID="588db7b9-6a66-4c02-be83-8766e47c211c" Dec 15 12:43:57 crc kubenswrapper[4719]: I1215 12:43:57.476965 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:43:57 crc kubenswrapper[4719]: E1215 12:43:57.477624 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:44:06 crc kubenswrapper[4719]: I1215 12:44:06.061866 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" event={"ID":"588db7b9-6a66-4c02-be83-8766e47c211c","Type":"ContainerStarted","Data":"57e117d08dcb6f25deede44f8f9218e4d2a4a67de0cc1f37ad656f59ec961145"} Dec 15 12:44:06 crc kubenswrapper[4719]: I1215 12:44:06.094799 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" podStartSLOduration=2.390239615 podStartE2EDuration="37.094781505s" podCreationTimestamp="2025-12-15 12:43:29 +0000 UTC" firstStartedPulling="2025-12-15 12:43:30.721003267 +0000 UTC m=+1571.663296297" lastFinishedPulling="2025-12-15 12:44:05.425545157 +0000 UTC m=+1606.367838187" observedRunningTime="2025-12-15 12:44:06.089360185 +0000 UTC m=+1607.031653225" watchObservedRunningTime="2025-12-15 12:44:06.094781505 +0000 UTC m=+1607.037074535" Dec 15 12:44:08 crc kubenswrapper[4719]: I1215 12:44:08.476665 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:44:08 crc kubenswrapper[4719]: E1215 12:44:08.478273 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:44:18 crc kubenswrapper[4719]: I1215 12:44:18.162776 4719 generic.go:334] "Generic (PLEG): container finished" podID="588db7b9-6a66-4c02-be83-8766e47c211c" containerID="57e117d08dcb6f25deede44f8f9218e4d2a4a67de0cc1f37ad656f59ec961145" exitCode=0 Dec 15 12:44:18 crc kubenswrapper[4719]: I1215 12:44:18.162873 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" event={"ID":"588db7b9-6a66-4c02-be83-8766e47c211c","Type":"ContainerDied","Data":"57e117d08dcb6f25deede44f8f9218e4d2a4a67de0cc1f37ad656f59ec961145"} Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.627458 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.689916 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzsrl\" (UniqueName: \"kubernetes.io/projected/588db7b9-6a66-4c02-be83-8766e47c211c-kube-api-access-pzsrl\") pod \"588db7b9-6a66-4c02-be83-8766e47c211c\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.689970 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-repo-setup-combined-ca-bundle\") pod \"588db7b9-6a66-4c02-be83-8766e47c211c\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.690020 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-inventory\") pod \"588db7b9-6a66-4c02-be83-8766e47c211c\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.690085 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-ssh-key\") pod \"588db7b9-6a66-4c02-be83-8766e47c211c\" (UID: \"588db7b9-6a66-4c02-be83-8766e47c211c\") " Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.698242 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "588db7b9-6a66-4c02-be83-8766e47c211c" (UID: "588db7b9-6a66-4c02-be83-8766e47c211c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.698363 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/588db7b9-6a66-4c02-be83-8766e47c211c-kube-api-access-pzsrl" (OuterVolumeSpecName: "kube-api-access-pzsrl") pod "588db7b9-6a66-4c02-be83-8766e47c211c" (UID: "588db7b9-6a66-4c02-be83-8766e47c211c"). InnerVolumeSpecName "kube-api-access-pzsrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.721937 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-inventory" (OuterVolumeSpecName: "inventory") pod "588db7b9-6a66-4c02-be83-8766e47c211c" (UID: "588db7b9-6a66-4c02-be83-8766e47c211c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.726388 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "588db7b9-6a66-4c02-be83-8766e47c211c" (UID: "588db7b9-6a66-4c02-be83-8766e47c211c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.792376 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzsrl\" (UniqueName: \"kubernetes.io/projected/588db7b9-6a66-4c02-be83-8766e47c211c-kube-api-access-pzsrl\") on node \"crc\" DevicePath \"\"" Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.792420 4719 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.792433 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:44:19 crc kubenswrapper[4719]: I1215 12:44:19.792443 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/588db7b9-6a66-4c02-be83-8766e47c211c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.214783 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" event={"ID":"588db7b9-6a66-4c02-be83-8766e47c211c","Type":"ContainerDied","Data":"358f83f9666dedc4a353f2562aefe6588190c939170afcdb7a1eddc1ff6cbaf6"} Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.214838 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="358f83f9666dedc4a353f2562aefe6588190c939170afcdb7a1eddc1ff6cbaf6" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.214844 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.299179 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt"] Dec 15 12:44:20 crc kubenswrapper[4719]: E1215 12:44:20.300108 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="588db7b9-6a66-4c02-be83-8766e47c211c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.300229 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="588db7b9-6a66-4c02-be83-8766e47c211c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.300491 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="588db7b9-6a66-4c02-be83-8766e47c211c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.301300 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.304569 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.307579 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.309774 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.309772 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.315476 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt"] Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.401791 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbvjq\" (UniqueName: \"kubernetes.io/projected/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-kube-api-access-nbvjq\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xjgnt\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.402117 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xjgnt\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.402278 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xjgnt\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.505199 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xjgnt\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.505344 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xjgnt\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.505629 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbvjq\" (UniqueName: \"kubernetes.io/projected/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-kube-api-access-nbvjq\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xjgnt\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.512691 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xjgnt\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.518715 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xjgnt\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.529557 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbvjq\" (UniqueName: \"kubernetes.io/projected/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-kube-api-access-nbvjq\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xjgnt\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:20 crc kubenswrapper[4719]: I1215 12:44:20.621711 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:21 crc kubenswrapper[4719]: I1215 12:44:21.154897 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt"] Dec 15 12:44:21 crc kubenswrapper[4719]: I1215 12:44:21.225896 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" event={"ID":"ed08a59f-ed8b-4195-b916-0d04c3e01f9f","Type":"ContainerStarted","Data":"e0e9a11d99908c874c707b70c04558ae726426efcaff745053706ffeef04b866"} Dec 15 12:44:21 crc kubenswrapper[4719]: I1215 12:44:21.476030 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:44:21 crc kubenswrapper[4719]: E1215 12:44:21.476476 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:44:22 crc kubenswrapper[4719]: I1215 12:44:22.234911 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" event={"ID":"ed08a59f-ed8b-4195-b916-0d04c3e01f9f","Type":"ContainerStarted","Data":"9dd9242c63b9eb00dfaae4126b2441cc2830a307888ccf3c67c49f7007383387"} Dec 15 12:44:22 crc kubenswrapper[4719]: I1215 12:44:22.252577 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" podStartSLOduration=1.646803347 podStartE2EDuration="2.252556743s" podCreationTimestamp="2025-12-15 12:44:20 +0000 UTC" firstStartedPulling="2025-12-15 12:44:21.170236537 +0000 UTC m=+1622.112529567" lastFinishedPulling="2025-12-15 12:44:21.775989933 +0000 UTC m=+1622.718282963" observedRunningTime="2025-12-15 12:44:22.251044185 +0000 UTC m=+1623.193337215" watchObservedRunningTime="2025-12-15 12:44:22.252556743 +0000 UTC m=+1623.194849783" Dec 15 12:44:25 crc kubenswrapper[4719]: I1215 12:44:25.270190 4719 generic.go:334] "Generic (PLEG): container finished" podID="ed08a59f-ed8b-4195-b916-0d04c3e01f9f" containerID="9dd9242c63b9eb00dfaae4126b2441cc2830a307888ccf3c67c49f7007383387" exitCode=0 Dec 15 12:44:25 crc kubenswrapper[4719]: I1215 12:44:25.270298 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" event={"ID":"ed08a59f-ed8b-4195-b916-0d04c3e01f9f","Type":"ContainerDied","Data":"9dd9242c63b9eb00dfaae4126b2441cc2830a307888ccf3c67c49f7007383387"} Dec 15 12:44:26 crc kubenswrapper[4719]: I1215 12:44:26.705086 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:26 crc kubenswrapper[4719]: I1215 12:44:26.714599 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-inventory\") pod \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " Dec 15 12:44:26 crc kubenswrapper[4719]: I1215 12:44:26.714678 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-ssh-key\") pod \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " Dec 15 12:44:26 crc kubenswrapper[4719]: I1215 12:44:26.714709 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbvjq\" (UniqueName: \"kubernetes.io/projected/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-kube-api-access-nbvjq\") pod \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\" (UID: \"ed08a59f-ed8b-4195-b916-0d04c3e01f9f\") " Dec 15 12:44:26 crc kubenswrapper[4719]: I1215 12:44:26.725988 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-kube-api-access-nbvjq" (OuterVolumeSpecName: "kube-api-access-nbvjq") pod "ed08a59f-ed8b-4195-b916-0d04c3e01f9f" (UID: "ed08a59f-ed8b-4195-b916-0d04c3e01f9f"). InnerVolumeSpecName "kube-api-access-nbvjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:44:26 crc kubenswrapper[4719]: I1215 12:44:26.747249 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ed08a59f-ed8b-4195-b916-0d04c3e01f9f" (UID: "ed08a59f-ed8b-4195-b916-0d04c3e01f9f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:44:26 crc kubenswrapper[4719]: I1215 12:44:26.777154 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-inventory" (OuterVolumeSpecName: "inventory") pod "ed08a59f-ed8b-4195-b916-0d04c3e01f9f" (UID: "ed08a59f-ed8b-4195-b916-0d04c3e01f9f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:44:26 crc kubenswrapper[4719]: I1215 12:44:26.816534 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:44:26 crc kubenswrapper[4719]: I1215 12:44:26.816593 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbvjq\" (UniqueName: \"kubernetes.io/projected/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-kube-api-access-nbvjq\") on node \"crc\" DevicePath \"\"" Dec 15 12:44:26 crc kubenswrapper[4719]: I1215 12:44:26.816605 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed08a59f-ed8b-4195-b916-0d04c3e01f9f-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.293140 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" event={"ID":"ed08a59f-ed8b-4195-b916-0d04c3e01f9f","Type":"ContainerDied","Data":"e0e9a11d99908c874c707b70c04558ae726426efcaff745053706ffeef04b866"} Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.293188 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0e9a11d99908c874c707b70c04558ae726426efcaff745053706ffeef04b866" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.293244 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xjgnt" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.361251 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w"] Dec 15 12:44:27 crc kubenswrapper[4719]: E1215 12:44:27.362024 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed08a59f-ed8b-4195-b916-0d04c3e01f9f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.362110 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed08a59f-ed8b-4195-b916-0d04c3e01f9f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.362393 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed08a59f-ed8b-4195-b916-0d04c3e01f9f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.363221 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.370307 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.370393 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.370485 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.370581 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.399272 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w"] Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.530180 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.530293 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.530366 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.530436 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rnrm\" (UniqueName: \"kubernetes.io/projected/19e3b1bb-a052-47fd-a833-bed95ef223fc-kube-api-access-4rnrm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.632291 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.632376 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rnrm\" (UniqueName: \"kubernetes.io/projected/19e3b1bb-a052-47fd-a833-bed95ef223fc-kube-api-access-4rnrm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.632543 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.632607 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.637354 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.637579 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.640589 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.647936 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rnrm\" (UniqueName: \"kubernetes.io/projected/19e3b1bb-a052-47fd-a833-bed95ef223fc-kube-api-access-4rnrm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:27 crc kubenswrapper[4719]: I1215 12:44:27.684910 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:44:28 crc kubenswrapper[4719]: I1215 12:44:28.206784 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w"] Dec 15 12:44:28 crc kubenswrapper[4719]: W1215 12:44:28.210112 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19e3b1bb_a052_47fd_a833_bed95ef223fc.slice/crio-38b652a6c3f7587227be0c8e00fbd5d39d3ce8c93599fb9e0b392a09c9906c15 WatchSource:0}: Error finding container 38b652a6c3f7587227be0c8e00fbd5d39d3ce8c93599fb9e0b392a09c9906c15: Status 404 returned error can't find the container with id 38b652a6c3f7587227be0c8e00fbd5d39d3ce8c93599fb9e0b392a09c9906c15 Dec 15 12:44:28 crc kubenswrapper[4719]: I1215 12:44:28.302596 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" event={"ID":"19e3b1bb-a052-47fd-a833-bed95ef223fc","Type":"ContainerStarted","Data":"38b652a6c3f7587227be0c8e00fbd5d39d3ce8c93599fb9e0b392a09c9906c15"} Dec 15 12:44:29 crc kubenswrapper[4719]: I1215 12:44:29.404997 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" event={"ID":"19e3b1bb-a052-47fd-a833-bed95ef223fc","Type":"ContainerStarted","Data":"41de04cd0b8137c018476f5528f5ef83832c18f34ba6a493bd456a80cc98d72c"} Dec 15 12:44:29 crc kubenswrapper[4719]: I1215 12:44:29.430761 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" podStartSLOduration=1.92571786 podStartE2EDuration="2.430745004s" podCreationTimestamp="2025-12-15 12:44:27 +0000 UTC" firstStartedPulling="2025-12-15 12:44:28.213138232 +0000 UTC m=+1629.155431262" lastFinishedPulling="2025-12-15 12:44:28.718165366 +0000 UTC m=+1629.660458406" observedRunningTime="2025-12-15 12:44:29.422939309 +0000 UTC m=+1630.365232349" watchObservedRunningTime="2025-12-15 12:44:29.430745004 +0000 UTC m=+1630.373038034" Dec 15 12:44:32 crc kubenswrapper[4719]: I1215 12:44:32.376997 4719 scope.go:117] "RemoveContainer" containerID="524d274eac51ccdfaa57b4292a48f01012b56b79adfbb9cfd91ef9a3471ee4eb" Dec 15 12:44:34 crc kubenswrapper[4719]: I1215 12:44:34.476509 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:44:34 crc kubenswrapper[4719]: E1215 12:44:34.477612 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:44:46 crc kubenswrapper[4719]: I1215 12:44:46.476499 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:44:46 crc kubenswrapper[4719]: E1215 12:44:46.477594 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:44:59 crc kubenswrapper[4719]: I1215 12:44:59.488662 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:44:59 crc kubenswrapper[4719]: E1215 12:44:59.489590 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.146442 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q"] Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.148636 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.151202 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.151253 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.161732 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q"] Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.294466 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4537bedf-a3f6-4c61-b53f-2438beff6c75-config-volume\") pod \"collect-profiles-29430045-v2g7q\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.294545 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4537bedf-a3f6-4c61-b53f-2438beff6c75-secret-volume\") pod \"collect-profiles-29430045-v2g7q\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.294695 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmgdn\" (UniqueName: \"kubernetes.io/projected/4537bedf-a3f6-4c61-b53f-2438beff6c75-kube-api-access-gmgdn\") pod \"collect-profiles-29430045-v2g7q\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.396883 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmgdn\" (UniqueName: \"kubernetes.io/projected/4537bedf-a3f6-4c61-b53f-2438beff6c75-kube-api-access-gmgdn\") pod \"collect-profiles-29430045-v2g7q\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.397223 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4537bedf-a3f6-4c61-b53f-2438beff6c75-config-volume\") pod \"collect-profiles-29430045-v2g7q\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.397377 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4537bedf-a3f6-4c61-b53f-2438beff6c75-secret-volume\") pod \"collect-profiles-29430045-v2g7q\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.398121 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4537bedf-a3f6-4c61-b53f-2438beff6c75-config-volume\") pod \"collect-profiles-29430045-v2g7q\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.417137 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4537bedf-a3f6-4c61-b53f-2438beff6c75-secret-volume\") pod \"collect-profiles-29430045-v2g7q\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.424515 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmgdn\" (UniqueName: \"kubernetes.io/projected/4537bedf-a3f6-4c61-b53f-2438beff6c75-kube-api-access-gmgdn\") pod \"collect-profiles-29430045-v2g7q\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.470174 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:00 crc kubenswrapper[4719]: I1215 12:45:00.927537 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q"] Dec 15 12:45:01 crc kubenswrapper[4719]: I1215 12:45:01.686700 4719 generic.go:334] "Generic (PLEG): container finished" podID="4537bedf-a3f6-4c61-b53f-2438beff6c75" containerID="685de466f1d4bfacb9a74afddc5a3277dd1b20873ba53fe5bd954fbc2ba1f3c4" exitCode=0 Dec 15 12:45:01 crc kubenswrapper[4719]: I1215 12:45:01.686804 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" event={"ID":"4537bedf-a3f6-4c61-b53f-2438beff6c75","Type":"ContainerDied","Data":"685de466f1d4bfacb9a74afddc5a3277dd1b20873ba53fe5bd954fbc2ba1f3c4"} Dec 15 12:45:01 crc kubenswrapper[4719]: I1215 12:45:01.687733 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" event={"ID":"4537bedf-a3f6-4c61-b53f-2438beff6c75","Type":"ContainerStarted","Data":"3cbf66318a40ce8a4e8ed40fa8d5fa9d466084ab5692bb6c2975f597dc243c3a"} Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.040542 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.152697 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmgdn\" (UniqueName: \"kubernetes.io/projected/4537bedf-a3f6-4c61-b53f-2438beff6c75-kube-api-access-gmgdn\") pod \"4537bedf-a3f6-4c61-b53f-2438beff6c75\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.152777 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4537bedf-a3f6-4c61-b53f-2438beff6c75-secret-volume\") pod \"4537bedf-a3f6-4c61-b53f-2438beff6c75\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.152804 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4537bedf-a3f6-4c61-b53f-2438beff6c75-config-volume\") pod \"4537bedf-a3f6-4c61-b53f-2438beff6c75\" (UID: \"4537bedf-a3f6-4c61-b53f-2438beff6c75\") " Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.153696 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4537bedf-a3f6-4c61-b53f-2438beff6c75-config-volume" (OuterVolumeSpecName: "config-volume") pod "4537bedf-a3f6-4c61-b53f-2438beff6c75" (UID: "4537bedf-a3f6-4c61-b53f-2438beff6c75"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.158413 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4537bedf-a3f6-4c61-b53f-2438beff6c75-kube-api-access-gmgdn" (OuterVolumeSpecName: "kube-api-access-gmgdn") pod "4537bedf-a3f6-4c61-b53f-2438beff6c75" (UID: "4537bedf-a3f6-4c61-b53f-2438beff6c75"). InnerVolumeSpecName "kube-api-access-gmgdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.159031 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4537bedf-a3f6-4c61-b53f-2438beff6c75-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4537bedf-a3f6-4c61-b53f-2438beff6c75" (UID: "4537bedf-a3f6-4c61-b53f-2438beff6c75"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.255816 4719 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4537bedf-a3f6-4c61-b53f-2438beff6c75-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.255880 4719 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4537bedf-a3f6-4c61-b53f-2438beff6c75-config-volume\") on node \"crc\" DevicePath \"\"" Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.255897 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmgdn\" (UniqueName: \"kubernetes.io/projected/4537bedf-a3f6-4c61-b53f-2438beff6c75-kube-api-access-gmgdn\") on node \"crc\" DevicePath \"\"" Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.705597 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" event={"ID":"4537bedf-a3f6-4c61-b53f-2438beff6c75","Type":"ContainerDied","Data":"3cbf66318a40ce8a4e8ed40fa8d5fa9d466084ab5692bb6c2975f597dc243c3a"} Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.705945 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cbf66318a40ce8a4e8ed40fa8d5fa9d466084ab5692bb6c2975f597dc243c3a" Dec 15 12:45:03 crc kubenswrapper[4719]: I1215 12:45:03.705668 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q" Dec 15 12:45:12 crc kubenswrapper[4719]: I1215 12:45:12.477448 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:45:12 crc kubenswrapper[4719]: E1215 12:45:12.479591 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:45:27 crc kubenswrapper[4719]: I1215 12:45:27.476263 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:45:27 crc kubenswrapper[4719]: E1215 12:45:27.477055 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:45:42 crc kubenswrapper[4719]: I1215 12:45:42.478147 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:45:42 crc kubenswrapper[4719]: E1215 12:45:42.483503 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:45:56 crc kubenswrapper[4719]: I1215 12:45:56.476957 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:45:56 crc kubenswrapper[4719]: E1215 12:45:56.478165 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:46:08 crc kubenswrapper[4719]: I1215 12:46:08.476596 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:46:08 crc kubenswrapper[4719]: E1215 12:46:08.477351 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:46:22 crc kubenswrapper[4719]: I1215 12:46:22.476323 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:46:22 crc kubenswrapper[4719]: E1215 12:46:22.477100 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:46:36 crc kubenswrapper[4719]: I1215 12:46:36.047577 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2498-account-create-update-bkr7v"] Dec 15 12:46:36 crc kubenswrapper[4719]: I1215 12:46:36.059249 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-9mxg6"] Dec 15 12:46:36 crc kubenswrapper[4719]: I1215 12:46:36.073652 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-00d1-account-create-update-tpf9v"] Dec 15 12:46:36 crc kubenswrapper[4719]: I1215 12:46:36.083768 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-5jgq2"] Dec 15 12:46:36 crc kubenswrapper[4719]: I1215 12:46:36.115899 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2498-account-create-update-bkr7v"] Dec 15 12:46:36 crc kubenswrapper[4719]: I1215 12:46:36.126845 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-00d1-account-create-update-tpf9v"] Dec 15 12:46:36 crc kubenswrapper[4719]: I1215 12:46:36.135548 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-9mxg6"] Dec 15 12:46:36 crc kubenswrapper[4719]: I1215 12:46:36.145497 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-5jgq2"] Dec 15 12:46:36 crc kubenswrapper[4719]: I1215 12:46:36.476903 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:46:36 crc kubenswrapper[4719]: E1215 12:46:36.477244 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:46:37 crc kubenswrapper[4719]: I1215 12:46:37.489937 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="392caf51-84a6-43d0-9bae-edbf23c1ef0a" path="/var/lib/kubelet/pods/392caf51-84a6-43d0-9bae-edbf23c1ef0a/volumes" Dec 15 12:46:37 crc kubenswrapper[4719]: I1215 12:46:37.491881 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50261891-12d4-4ec9-a0c0-e3b1d07859b6" path="/var/lib/kubelet/pods/50261891-12d4-4ec9-a0c0-e3b1d07859b6/volumes" Dec 15 12:46:37 crc kubenswrapper[4719]: I1215 12:46:37.493463 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="709db400-dea1-4f84-a1bd-e3c6bdc44644" path="/var/lib/kubelet/pods/709db400-dea1-4f84-a1bd-e3c6bdc44644/volumes" Dec 15 12:46:37 crc kubenswrapper[4719]: I1215 12:46:37.495385 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5121ea0-ad37-471e-9eca-d88c97375e25" path="/var/lib/kubelet/pods/e5121ea0-ad37-471e-9eca-d88c97375e25/volumes" Dec 15 12:46:41 crc kubenswrapper[4719]: I1215 12:46:41.033818 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-dndhn"] Dec 15 12:46:41 crc kubenswrapper[4719]: I1215 12:46:41.044185 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-2228-account-create-update-rwhbn"] Dec 15 12:46:41 crc kubenswrapper[4719]: I1215 12:46:41.053998 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-dndhn"] Dec 15 12:46:41 crc kubenswrapper[4719]: I1215 12:46:41.062677 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-2228-account-create-update-rwhbn"] Dec 15 12:46:41 crc kubenswrapper[4719]: I1215 12:46:41.498687 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7238f625-5678-487a-91fc-a3a7686441bd" path="/var/lib/kubelet/pods/7238f625-5678-487a-91fc-a3a7686441bd/volumes" Dec 15 12:46:41 crc kubenswrapper[4719]: I1215 12:46:41.501139 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaaec482-1c98-40c9-a402-97e6b8241a98" path="/var/lib/kubelet/pods/eaaec482-1c98-40c9-a402-97e6b8241a98/volumes" Dec 15 12:46:51 crc kubenswrapper[4719]: I1215 12:46:51.475479 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:46:51 crc kubenswrapper[4719]: E1215 12:46:51.476203 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:47:04 crc kubenswrapper[4719]: I1215 12:47:04.476749 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:47:04 crc kubenswrapper[4719]: E1215 12:47:04.477877 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:47:12 crc kubenswrapper[4719]: I1215 12:47:12.039675 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-nf7qt"] Dec 15 12:47:12 crc kubenswrapper[4719]: I1215 12:47:12.052044 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-nf7qt"] Dec 15 12:47:13 crc kubenswrapper[4719]: I1215 12:47:13.502387 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="479f222a-bfc2-4257-9588-941538d75171" path="/var/lib/kubelet/pods/479f222a-bfc2-4257-9588-941538d75171/volumes" Dec 15 12:47:18 crc kubenswrapper[4719]: I1215 12:47:18.477423 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:47:18 crc kubenswrapper[4719]: E1215 12:47:18.478563 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:47:24 crc kubenswrapper[4719]: I1215 12:47:24.088001 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-lgf62"] Dec 15 12:47:24 crc kubenswrapper[4719]: I1215 12:47:24.106000 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-539a-account-create-update-zqll7"] Dec 15 12:47:24 crc kubenswrapper[4719]: I1215 12:47:24.116473 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-lgf62"] Dec 15 12:47:24 crc kubenswrapper[4719]: I1215 12:47:24.125284 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-539a-account-create-update-zqll7"] Dec 15 12:47:24 crc kubenswrapper[4719]: I1215 12:47:24.132700 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-6f1e-account-create-update-bps6c"] Dec 15 12:47:24 crc kubenswrapper[4719]: I1215 12:47:24.140248 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d2d3-account-create-update-8lpdk"] Dec 15 12:47:24 crc kubenswrapper[4719]: I1215 12:47:24.147465 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-6f1e-account-create-update-bps6c"] Dec 15 12:47:24 crc kubenswrapper[4719]: I1215 12:47:24.155407 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d2d3-account-create-update-8lpdk"] Dec 15 12:47:25 crc kubenswrapper[4719]: I1215 12:47:25.029056 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-tcmf2"] Dec 15 12:47:25 crc kubenswrapper[4719]: I1215 12:47:25.041068 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-75qg9"] Dec 15 12:47:25 crc kubenswrapper[4719]: I1215 12:47:25.049708 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-75qg9"] Dec 15 12:47:25 crc kubenswrapper[4719]: I1215 12:47:25.083591 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-tcmf2"] Dec 15 12:47:25 crc kubenswrapper[4719]: I1215 12:47:25.488139 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0381936c-f528-4662-8a86-558c341065c0" path="/var/lib/kubelet/pods/0381936c-f528-4662-8a86-558c341065c0/volumes" Dec 15 12:47:25 crc kubenswrapper[4719]: I1215 12:47:25.491344 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83819057-000b-4f9e-b8fc-3478bc403f4e" path="/var/lib/kubelet/pods/83819057-000b-4f9e-b8fc-3478bc403f4e/volumes" Dec 15 12:47:25 crc kubenswrapper[4719]: I1215 12:47:25.492909 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0e6d128-ab9b-4a15-bef7-c986942dc16f" path="/var/lib/kubelet/pods/c0e6d128-ab9b-4a15-bef7-c986942dc16f/volumes" Dec 15 12:47:25 crc kubenswrapper[4719]: I1215 12:47:25.493956 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d04b79-2fb3-44a2-9b61-d6ab75ad9752" path="/var/lib/kubelet/pods/d8d04b79-2fb3-44a2-9b61-d6ab75ad9752/volumes" Dec 15 12:47:25 crc kubenswrapper[4719]: I1215 12:47:25.495652 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2ec6917-fd83-4b5b-b904-6a0e4d453cf1" path="/var/lib/kubelet/pods/e2ec6917-fd83-4b5b-b904-6a0e4d453cf1/volumes" Dec 15 12:47:25 crc kubenswrapper[4719]: I1215 12:47:25.497110 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdc568d9-e75c-4f9b-a426-4dbc175a9f44" path="/var/lib/kubelet/pods/fdc568d9-e75c-4f9b-a426-4dbc175a9f44/volumes" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.508173 4719 scope.go:117] "RemoveContainer" containerID="293b2bef238e11454bddbe8d8d1e8bfab6dea3f513dcaac1bfe8995a320a3589" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.538346 4719 scope.go:117] "RemoveContainer" containerID="fb301aeb8662ca386a1b946e9d89d0d5126b0c82f9f845b0c7de32fa368ebbca" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.585876 4719 scope.go:117] "RemoveContainer" containerID="2c9715a195700a9e7cd49c8aea13f28966c101cf2b45bc29b3a0470cb83f8e53" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.635993 4719 scope.go:117] "RemoveContainer" containerID="9d0f821426f732ad4b5c673b55eff4ecbedba7d99257789e357e7377e14f941f" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.692705 4719 scope.go:117] "RemoveContainer" containerID="96c75bdc1d8541db7b6e7a92f093ee49a969e8d1d327ae74d40290660ce7c297" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.724741 4719 scope.go:117] "RemoveContainer" containerID="b8e610d478d574b84ff4a581bf6527c98612b4555494e76fce65dfac02699355" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.768067 4719 scope.go:117] "RemoveContainer" containerID="530d8a4b0fa480da3faa6731792800736b86d48a1107001a800489dc05bafb7b" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.796828 4719 scope.go:117] "RemoveContainer" containerID="6e1db7d818f96e2c02ed751e956127418fd37244d77f05c42873cad1ce04282d" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.821631 4719 scope.go:117] "RemoveContainer" containerID="f8763d2f1ab31a94b363b104c80cc7d11a71d7999d4b244c60be71c819c74010" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.846605 4719 scope.go:117] "RemoveContainer" containerID="5f094fe3136a3f3ce0225bc95beac5be77d50f6bddf8229c8d1be2f03cab6b51" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.871295 4719 scope.go:117] "RemoveContainer" containerID="e5cc5f5f146af9ea8b7ade3b01b578bb9e1e5f2762bf9bfb445a5a001521bd04" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.893496 4719 scope.go:117] "RemoveContainer" containerID="3616a57fe36a9e7aea2d305291ee91e9b90b8da3223af440ba8b48459410b4ec" Dec 15 12:47:32 crc kubenswrapper[4719]: I1215 12:47:32.913414 4719 scope.go:117] "RemoveContainer" containerID="e9d54eb7401cd1368de91a626037d8fcc0ae6d7646d63d3a8c03402a94969a37" Dec 15 12:47:33 crc kubenswrapper[4719]: I1215 12:47:33.476190 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:47:33 crc kubenswrapper[4719]: E1215 12:47:33.476510 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:47:37 crc kubenswrapper[4719]: I1215 12:47:37.040174 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-c6wxw"] Dec 15 12:47:37 crc kubenswrapper[4719]: I1215 12:47:37.053495 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-c6wxw"] Dec 15 12:47:37 crc kubenswrapper[4719]: I1215 12:47:37.495738 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c64d8b6-1005-484e-b068-95d462167a8c" path="/var/lib/kubelet/pods/4c64d8b6-1005-484e-b068-95d462167a8c/volumes" Dec 15 12:47:44 crc kubenswrapper[4719]: I1215 12:47:44.476576 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:47:44 crc kubenswrapper[4719]: E1215 12:47:44.477547 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:47:55 crc kubenswrapper[4719]: I1215 12:47:55.478080 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:47:55 crc kubenswrapper[4719]: E1215 12:47:55.479092 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:48:07 crc kubenswrapper[4719]: I1215 12:48:07.476171 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:48:07 crc kubenswrapper[4719]: E1215 12:48:07.477323 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:48:14 crc kubenswrapper[4719]: I1215 12:48:14.187201 4719 generic.go:334] "Generic (PLEG): container finished" podID="19e3b1bb-a052-47fd-a833-bed95ef223fc" containerID="41de04cd0b8137c018476f5528f5ef83832c18f34ba6a493bd456a80cc98d72c" exitCode=0 Dec 15 12:48:14 crc kubenswrapper[4719]: I1215 12:48:14.187307 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" event={"ID":"19e3b1bb-a052-47fd-a833-bed95ef223fc","Type":"ContainerDied","Data":"41de04cd0b8137c018476f5528f5ef83832c18f34ba6a493bd456a80cc98d72c"} Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.694576 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.801148 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-ssh-key\") pod \"19e3b1bb-a052-47fd-a833-bed95ef223fc\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.801457 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rnrm\" (UniqueName: \"kubernetes.io/projected/19e3b1bb-a052-47fd-a833-bed95ef223fc-kube-api-access-4rnrm\") pod \"19e3b1bb-a052-47fd-a833-bed95ef223fc\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.801535 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-inventory\") pod \"19e3b1bb-a052-47fd-a833-bed95ef223fc\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.801688 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-bootstrap-combined-ca-bundle\") pod \"19e3b1bb-a052-47fd-a833-bed95ef223fc\" (UID: \"19e3b1bb-a052-47fd-a833-bed95ef223fc\") " Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.807018 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19e3b1bb-a052-47fd-a833-bed95ef223fc-kube-api-access-4rnrm" (OuterVolumeSpecName: "kube-api-access-4rnrm") pod "19e3b1bb-a052-47fd-a833-bed95ef223fc" (UID: "19e3b1bb-a052-47fd-a833-bed95ef223fc"). InnerVolumeSpecName "kube-api-access-4rnrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.811472 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "19e3b1bb-a052-47fd-a833-bed95ef223fc" (UID: "19e3b1bb-a052-47fd-a833-bed95ef223fc"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.832202 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-inventory" (OuterVolumeSpecName: "inventory") pod "19e3b1bb-a052-47fd-a833-bed95ef223fc" (UID: "19e3b1bb-a052-47fd-a833-bed95ef223fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.846016 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "19e3b1bb-a052-47fd-a833-bed95ef223fc" (UID: "19e3b1bb-a052-47fd-a833-bed95ef223fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.903777 4719 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.903810 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.903822 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rnrm\" (UniqueName: \"kubernetes.io/projected/19e3b1bb-a052-47fd-a833-bed95ef223fc-kube-api-access-4rnrm\") on node \"crc\" DevicePath \"\"" Dec 15 12:48:15 crc kubenswrapper[4719]: I1215 12:48:15.903832 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19e3b1bb-a052-47fd-a833-bed95ef223fc-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.207464 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" event={"ID":"19e3b1bb-a052-47fd-a833-bed95ef223fc","Type":"ContainerDied","Data":"38b652a6c3f7587227be0c8e00fbd5d39d3ce8c93599fb9e0b392a09c9906c15"} Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.207513 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38b652a6c3f7587227be0c8e00fbd5d39d3ce8c93599fb9e0b392a09c9906c15" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.207535 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.352236 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct"] Dec 15 12:48:16 crc kubenswrapper[4719]: E1215 12:48:16.352681 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4537bedf-a3f6-4c61-b53f-2438beff6c75" containerName="collect-profiles" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.352703 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="4537bedf-a3f6-4c61-b53f-2438beff6c75" containerName="collect-profiles" Dec 15 12:48:16 crc kubenswrapper[4719]: E1215 12:48:16.352727 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19e3b1bb-a052-47fd-a833-bed95ef223fc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.352736 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="19e3b1bb-a052-47fd-a833-bed95ef223fc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.353011 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="4537bedf-a3f6-4c61-b53f-2438beff6c75" containerName="collect-profiles" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.353039 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="19e3b1bb-a052-47fd-a833-bed95ef223fc" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.353633 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.356340 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.356765 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.356989 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.358480 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.365138 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct"] Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.514671 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.515061 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.515242 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc5xq\" (UniqueName: \"kubernetes.io/projected/b72d7eb2-0c87-4377-8338-118b53d43241-kube-api-access-dc5xq\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.616944 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.617190 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.617338 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc5xq\" (UniqueName: \"kubernetes.io/projected/b72d7eb2-0c87-4377-8338-118b53d43241-kube-api-access-dc5xq\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.623653 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.624101 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.691722 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc5xq\" (UniqueName: \"kubernetes.io/projected/b72d7eb2-0c87-4377-8338-118b53d43241-kube-api-access-dc5xq\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:16 crc kubenswrapper[4719]: I1215 12:48:16.693549 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:48:17 crc kubenswrapper[4719]: I1215 12:48:17.256398 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct"] Dec 15 12:48:17 crc kubenswrapper[4719]: I1215 12:48:17.270084 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 12:48:18 crc kubenswrapper[4719]: I1215 12:48:18.230421 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" event={"ID":"b72d7eb2-0c87-4377-8338-118b53d43241","Type":"ContainerStarted","Data":"7d6afa53ab47f91e78f7a8ef437058c3fec80a7b472ca2f263415663ada46330"} Dec 15 12:48:20 crc kubenswrapper[4719]: I1215 12:48:20.246498 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" event={"ID":"b72d7eb2-0c87-4377-8338-118b53d43241","Type":"ContainerStarted","Data":"47e5fc4d08c8588b87b0e5cb9b3b7329cce9e18f226ae8de84d6b51f9ebdc469"} Dec 15 12:48:20 crc kubenswrapper[4719]: I1215 12:48:20.264642 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" podStartSLOduration=2.114139805 podStartE2EDuration="4.264623917s" podCreationTimestamp="2025-12-15 12:48:16 +0000 UTC" firstStartedPulling="2025-12-15 12:48:17.269565665 +0000 UTC m=+1858.211858705" lastFinishedPulling="2025-12-15 12:48:19.420049747 +0000 UTC m=+1860.362342817" observedRunningTime="2025-12-15 12:48:20.259162235 +0000 UTC m=+1861.201455275" watchObservedRunningTime="2025-12-15 12:48:20.264623917 +0000 UTC m=+1861.206916947" Dec 15 12:48:22 crc kubenswrapper[4719]: I1215 12:48:22.478014 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:48:23 crc kubenswrapper[4719]: I1215 12:48:23.271638 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"23da24f9e77dd91d08a2fcb3b8667d259f3c0a3ecdb0953729476936101d7499"} Dec 15 12:48:33 crc kubenswrapper[4719]: I1215 12:48:33.190275 4719 scope.go:117] "RemoveContainer" containerID="d8fc87b95d5a2eaed35562d83d92e0bdefc4ed210c7b6338f3037671ab1736f6" Dec 15 12:48:34 crc kubenswrapper[4719]: I1215 12:48:34.065413 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-88rcn"] Dec 15 12:48:34 crc kubenswrapper[4719]: I1215 12:48:34.083538 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-88rcn"] Dec 15 12:48:35 crc kubenswrapper[4719]: I1215 12:48:35.026053 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nd852"] Dec 15 12:48:35 crc kubenswrapper[4719]: I1215 12:48:35.033547 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nd852"] Dec 15 12:48:35 crc kubenswrapper[4719]: I1215 12:48:35.491987 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="978132fa-fca3-4665-b133-682da5d02f93" path="/var/lib/kubelet/pods/978132fa-fca3-4665-b133-682da5d02f93/volumes" Dec 15 12:48:35 crc kubenswrapper[4719]: I1215 12:48:35.492706 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cba0d08d-bba6-4a68-bc44-c4cf1619ad4c" path="/var/lib/kubelet/pods/cba0d08d-bba6-4a68-bc44-c4cf1619ad4c/volumes" Dec 15 12:48:41 crc kubenswrapper[4719]: I1215 12:48:41.773379 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vhxfj"] Dec 15 12:48:41 crc kubenswrapper[4719]: I1215 12:48:41.775850 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:41 crc kubenswrapper[4719]: I1215 12:48:41.781939 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhxfj"] Dec 15 12:48:41 crc kubenswrapper[4719]: I1215 12:48:41.925679 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-catalog-content\") pod \"redhat-marketplace-vhxfj\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:41 crc kubenswrapper[4719]: I1215 12:48:41.925750 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b49wl\" (UniqueName: \"kubernetes.io/projected/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-kube-api-access-b49wl\") pod \"redhat-marketplace-vhxfj\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:41 crc kubenswrapper[4719]: I1215 12:48:41.925781 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-utilities\") pod \"redhat-marketplace-vhxfj\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:42 crc kubenswrapper[4719]: I1215 12:48:42.027008 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b49wl\" (UniqueName: \"kubernetes.io/projected/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-kube-api-access-b49wl\") pod \"redhat-marketplace-vhxfj\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:42 crc kubenswrapper[4719]: I1215 12:48:42.027296 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-utilities\") pod \"redhat-marketplace-vhxfj\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:42 crc kubenswrapper[4719]: I1215 12:48:42.027551 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-catalog-content\") pod \"redhat-marketplace-vhxfj\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:42 crc kubenswrapper[4719]: I1215 12:48:42.027773 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-utilities\") pod \"redhat-marketplace-vhxfj\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:42 crc kubenswrapper[4719]: I1215 12:48:42.028008 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-catalog-content\") pod \"redhat-marketplace-vhxfj\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:42 crc kubenswrapper[4719]: I1215 12:48:42.047729 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b49wl\" (UniqueName: \"kubernetes.io/projected/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-kube-api-access-b49wl\") pod \"redhat-marketplace-vhxfj\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:42 crc kubenswrapper[4719]: I1215 12:48:42.113596 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:42 crc kubenswrapper[4719]: W1215 12:48:42.566641 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ec135af_9bb6_45ff_9d75_4d226e9ed6bf.slice/crio-efd342020000a6abdcd845841697682274acb95742125e075ce1d95fb26afb1b WatchSource:0}: Error finding container efd342020000a6abdcd845841697682274acb95742125e075ce1d95fb26afb1b: Status 404 returned error can't find the container with id efd342020000a6abdcd845841697682274acb95742125e075ce1d95fb26afb1b Dec 15 12:48:42 crc kubenswrapper[4719]: I1215 12:48:42.572784 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhxfj"] Dec 15 12:48:43 crc kubenswrapper[4719]: I1215 12:48:43.481426 4719 generic.go:334] "Generic (PLEG): container finished" podID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerID="b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a" exitCode=0 Dec 15 12:48:43 crc kubenswrapper[4719]: I1215 12:48:43.495018 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhxfj" event={"ID":"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf","Type":"ContainerDied","Data":"b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a"} Dec 15 12:48:43 crc kubenswrapper[4719]: I1215 12:48:43.495067 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhxfj" event={"ID":"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf","Type":"ContainerStarted","Data":"efd342020000a6abdcd845841697682274acb95742125e075ce1d95fb26afb1b"} Dec 15 12:48:45 crc kubenswrapper[4719]: I1215 12:48:45.503582 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhxfj" event={"ID":"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf","Type":"ContainerStarted","Data":"e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e"} Dec 15 12:48:46 crc kubenswrapper[4719]: I1215 12:48:46.031898 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-8qr2l"] Dec 15 12:48:46 crc kubenswrapper[4719]: I1215 12:48:46.037682 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-8qr2l"] Dec 15 12:48:46 crc kubenswrapper[4719]: I1215 12:48:46.514046 4719 generic.go:334] "Generic (PLEG): container finished" podID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerID="e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e" exitCode=0 Dec 15 12:48:46 crc kubenswrapper[4719]: I1215 12:48:46.514091 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhxfj" event={"ID":"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf","Type":"ContainerDied","Data":"e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e"} Dec 15 12:48:47 crc kubenswrapper[4719]: I1215 12:48:47.488518 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58fea5e6-9ff9-4a44-be8a-311735af3e4b" path="/var/lib/kubelet/pods/58fea5e6-9ff9-4a44-be8a-311735af3e4b/volumes" Dec 15 12:48:47 crc kubenswrapper[4719]: I1215 12:48:47.522829 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhxfj" event={"ID":"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf","Type":"ContainerStarted","Data":"7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e"} Dec 15 12:48:47 crc kubenswrapper[4719]: I1215 12:48:47.550722 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vhxfj" podStartSLOduration=2.718163241 podStartE2EDuration="6.550704891s" podCreationTimestamp="2025-12-15 12:48:41 +0000 UTC" firstStartedPulling="2025-12-15 12:48:43.484290524 +0000 UTC m=+1884.426583554" lastFinishedPulling="2025-12-15 12:48:47.316832184 +0000 UTC m=+1888.259125204" observedRunningTime="2025-12-15 12:48:47.54403019 +0000 UTC m=+1888.486323220" watchObservedRunningTime="2025-12-15 12:48:47.550704891 +0000 UTC m=+1888.492997921" Dec 15 12:48:48 crc kubenswrapper[4719]: I1215 12:48:48.031348 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-9vwxh"] Dec 15 12:48:48 crc kubenswrapper[4719]: I1215 12:48:48.039805 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-9vwxh"] Dec 15 12:48:49 crc kubenswrapper[4719]: I1215 12:48:49.488211 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3243ab84-0a4c-4c74-83de-693c1c96e89e" path="/var/lib/kubelet/pods/3243ab84-0a4c-4c74-83de-693c1c96e89e/volumes" Dec 15 12:48:52 crc kubenswrapper[4719]: I1215 12:48:52.042985 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-gprbv"] Dec 15 12:48:52 crc kubenswrapper[4719]: I1215 12:48:52.053786 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-gprbv"] Dec 15 12:48:52 crc kubenswrapper[4719]: I1215 12:48:52.114309 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:52 crc kubenswrapper[4719]: I1215 12:48:52.114343 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:52 crc kubenswrapper[4719]: I1215 12:48:52.164935 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:52 crc kubenswrapper[4719]: I1215 12:48:52.631140 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:52 crc kubenswrapper[4719]: I1215 12:48:52.694570 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhxfj"] Dec 15 12:48:53 crc kubenswrapper[4719]: I1215 12:48:53.491308 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef9ed6e9-3050-4335-8ffc-97f889eb63e6" path="/var/lib/kubelet/pods/ef9ed6e9-3050-4335-8ffc-97f889eb63e6/volumes" Dec 15 12:48:54 crc kubenswrapper[4719]: I1215 12:48:54.590448 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vhxfj" podUID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerName="registry-server" containerID="cri-o://7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e" gracePeriod=2 Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.029621 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.174562 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-catalog-content\") pod \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.174824 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b49wl\" (UniqueName: \"kubernetes.io/projected/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-kube-api-access-b49wl\") pod \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.176587 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-utilities\") pod \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\" (UID: \"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf\") " Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.177968 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-utilities" (OuterVolumeSpecName: "utilities") pod "0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" (UID: "0ec135af-9bb6-45ff-9d75-4d226e9ed6bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.178839 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.190216 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-kube-api-access-b49wl" (OuterVolumeSpecName: "kube-api-access-b49wl") pod "0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" (UID: "0ec135af-9bb6-45ff-9d75-4d226e9ed6bf"). InnerVolumeSpecName "kube-api-access-b49wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.198139 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" (UID: "0ec135af-9bb6-45ff-9d75-4d226e9ed6bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.280045 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.280079 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b49wl\" (UniqueName: \"kubernetes.io/projected/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf-kube-api-access-b49wl\") on node \"crc\" DevicePath \"\"" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.602504 4719 generic.go:334] "Generic (PLEG): container finished" podID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerID="7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e" exitCode=0 Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.602593 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vhxfj" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.602587 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhxfj" event={"ID":"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf","Type":"ContainerDied","Data":"7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e"} Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.602941 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vhxfj" event={"ID":"0ec135af-9bb6-45ff-9d75-4d226e9ed6bf","Type":"ContainerDied","Data":"efd342020000a6abdcd845841697682274acb95742125e075ce1d95fb26afb1b"} Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.602969 4719 scope.go:117] "RemoveContainer" containerID="7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.629993 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhxfj"] Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.631157 4719 scope.go:117] "RemoveContainer" containerID="e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.638566 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vhxfj"] Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.652554 4719 scope.go:117] "RemoveContainer" containerID="b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.694392 4719 scope.go:117] "RemoveContainer" containerID="7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e" Dec 15 12:48:55 crc kubenswrapper[4719]: E1215 12:48:55.694840 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e\": container with ID starting with 7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e not found: ID does not exist" containerID="7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.694902 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e"} err="failed to get container status \"7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e\": rpc error: code = NotFound desc = could not find container \"7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e\": container with ID starting with 7e78d3111b32941d313c96624f99ee5fc176d951d8bf0d07e335cc5d43e7121e not found: ID does not exist" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.694926 4719 scope.go:117] "RemoveContainer" containerID="e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e" Dec 15 12:48:55 crc kubenswrapper[4719]: E1215 12:48:55.695549 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e\": container with ID starting with e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e not found: ID does not exist" containerID="e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.695569 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e"} err="failed to get container status \"e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e\": rpc error: code = NotFound desc = could not find container \"e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e\": container with ID starting with e5111baafbb287458cd7f7461077b949e0853ae649bb8900db77a7c440eab87e not found: ID does not exist" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.695582 4719 scope.go:117] "RemoveContainer" containerID="b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a" Dec 15 12:48:55 crc kubenswrapper[4719]: E1215 12:48:55.695893 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a\": container with ID starting with b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a not found: ID does not exist" containerID="b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a" Dec 15 12:48:55 crc kubenswrapper[4719]: I1215 12:48:55.695936 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a"} err="failed to get container status \"b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a\": rpc error: code = NotFound desc = could not find container \"b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a\": container with ID starting with b743e533d1e3ea1052ba245c2d87f4a9db173c6a782b1637b7b961b89b0dcd1a not found: ID does not exist" Dec 15 12:48:57 crc kubenswrapper[4719]: I1215 12:48:57.485314 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" path="/var/lib/kubelet/pods/0ec135af-9bb6-45ff-9d75-4d226e9ed6bf/volumes" Dec 15 12:49:33 crc kubenswrapper[4719]: I1215 12:49:33.290536 4719 scope.go:117] "RemoveContainer" containerID="68928debcaf4ded9eda72c222dfdb8418b8bf08b43ca2b04dfd19ea947b50a35" Dec 15 12:49:33 crc kubenswrapper[4719]: I1215 12:49:33.335893 4719 scope.go:117] "RemoveContainer" containerID="700804ea29d99bd4f42fffec72e1606580a670faee05531a7f62112e644c7a84" Dec 15 12:49:33 crc kubenswrapper[4719]: I1215 12:49:33.386959 4719 scope.go:117] "RemoveContainer" containerID="f2b161a88efddcd95d6d9476aa89c6ab52a625caa07795376362d6a570e0cdf2" Dec 15 12:49:33 crc kubenswrapper[4719]: I1215 12:49:33.439334 4719 scope.go:117] "RemoveContainer" containerID="f9a10627eba3e330e4bc0909a38d7d0e9a3da28f30c8c1e92e26e0c385194bd6" Dec 15 12:49:33 crc kubenswrapper[4719]: I1215 12:49:33.482171 4719 scope.go:117] "RemoveContainer" containerID="f78a6395ccee6de4c3c5dec602c3cb6643a39d904cb1023542b07be5867f3e20" Dec 15 12:50:12 crc kubenswrapper[4719]: I1215 12:50:12.043436 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-j88sc"] Dec 15 12:50:12 crc kubenswrapper[4719]: I1215 12:50:12.052309 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-j88sc"] Dec 15 12:50:12 crc kubenswrapper[4719]: I1215 12:50:12.330689 4719 generic.go:334] "Generic (PLEG): container finished" podID="b72d7eb2-0c87-4377-8338-118b53d43241" containerID="47e5fc4d08c8588b87b0e5cb9b3b7329cce9e18f226ae8de84d6b51f9ebdc469" exitCode=0 Dec 15 12:50:12 crc kubenswrapper[4719]: I1215 12:50:12.330739 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" event={"ID":"b72d7eb2-0c87-4377-8338-118b53d43241","Type":"ContainerDied","Data":"47e5fc4d08c8588b87b0e5cb9b3b7329cce9e18f226ae8de84d6b51f9ebdc469"} Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.040458 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-30e9-account-create-update-rtsbx"] Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.051998 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-30e9-account-create-update-rtsbx"] Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.062231 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-9663-account-create-update-fcjsm"] Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.072606 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-4957-account-create-update-r5lcc"] Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.082788 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-9663-account-create-update-fcjsm"] Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.091995 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-5w2hw"] Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.099818 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-bl7mv"] Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.107781 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-5w2hw"] Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.115569 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-4957-account-create-update-r5lcc"] Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.122324 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-bl7mv"] Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.489170 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30b90f8c-d014-4952-b250-1aaaa52ea020" path="/var/lib/kubelet/pods/30b90f8c-d014-4952-b250-1aaaa52ea020/volumes" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.489842 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42a85d8b-aca5-4aea-acfd-ee3ff752d272" path="/var/lib/kubelet/pods/42a85d8b-aca5-4aea-acfd-ee3ff752d272/volumes" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.490469 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73c6a473-c59b-481d-9047-a4006471710a" path="/var/lib/kubelet/pods/73c6a473-c59b-481d-9047-a4006471710a/volumes" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.491062 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5414c29-3e89-4a89-860d-9798ac533d81" path="/var/lib/kubelet/pods/b5414c29-3e89-4a89-860d-9798ac533d81/volumes" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.492314 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e200319d-f1b8-421b-ba02-a9613dba7363" path="/var/lib/kubelet/pods/e200319d-f1b8-421b-ba02-a9613dba7363/volumes" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.492960 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef50d20e-737c-486b-9759-08e17f34ce83" path="/var/lib/kubelet/pods/ef50d20e-737c-486b-9759-08e17f34ce83/volumes" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.733663 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.841489 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-ssh-key\") pod \"b72d7eb2-0c87-4377-8338-118b53d43241\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.841554 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc5xq\" (UniqueName: \"kubernetes.io/projected/b72d7eb2-0c87-4377-8338-118b53d43241-kube-api-access-dc5xq\") pod \"b72d7eb2-0c87-4377-8338-118b53d43241\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.841725 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-inventory\") pod \"b72d7eb2-0c87-4377-8338-118b53d43241\" (UID: \"b72d7eb2-0c87-4377-8338-118b53d43241\") " Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.848061 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b72d7eb2-0c87-4377-8338-118b53d43241-kube-api-access-dc5xq" (OuterVolumeSpecName: "kube-api-access-dc5xq") pod "b72d7eb2-0c87-4377-8338-118b53d43241" (UID: "b72d7eb2-0c87-4377-8338-118b53d43241"). InnerVolumeSpecName "kube-api-access-dc5xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.875393 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-inventory" (OuterVolumeSpecName: "inventory") pod "b72d7eb2-0c87-4377-8338-118b53d43241" (UID: "b72d7eb2-0c87-4377-8338-118b53d43241"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.882683 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b72d7eb2-0c87-4377-8338-118b53d43241" (UID: "b72d7eb2-0c87-4377-8338-118b53d43241"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.943489 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.943522 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc5xq\" (UniqueName: \"kubernetes.io/projected/b72d7eb2-0c87-4377-8338-118b53d43241-kube-api-access-dc5xq\") on node \"crc\" DevicePath \"\"" Dec 15 12:50:13 crc kubenswrapper[4719]: I1215 12:50:13.943534 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b72d7eb2-0c87-4377-8338-118b53d43241-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.351153 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" event={"ID":"b72d7eb2-0c87-4377-8338-118b53d43241","Type":"ContainerDied","Data":"7d6afa53ab47f91e78f7a8ef437058c3fec80a7b472ca2f263415663ada46330"} Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.351211 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d6afa53ab47f91e78f7a8ef437058c3fec80a7b472ca2f263415663ada46330" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.351244 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.471963 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z"] Dec 15 12:50:14 crc kubenswrapper[4719]: E1215 12:50:14.472401 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72d7eb2-0c87-4377-8338-118b53d43241" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.472422 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72d7eb2-0c87-4377-8338-118b53d43241" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 15 12:50:14 crc kubenswrapper[4719]: E1215 12:50:14.472446 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerName="extract-content" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.472455 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerName="extract-content" Dec 15 12:50:14 crc kubenswrapper[4719]: E1215 12:50:14.472470 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerName="registry-server" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.472478 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerName="registry-server" Dec 15 12:50:14 crc kubenswrapper[4719]: E1215 12:50:14.472492 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerName="extract-utilities" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.472500 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerName="extract-utilities" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.472716 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ec135af-9bb6-45ff-9d75-4d226e9ed6bf" containerName="registry-server" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.472750 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="b72d7eb2-0c87-4377-8338-118b53d43241" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.473603 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.475581 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.476034 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.476256 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.476423 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.495351 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z"] Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.555313 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7q45z\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.555947 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7q45z\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.556668 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgcfp\" (UniqueName: \"kubernetes.io/projected/2b3c8271-0d1c-4da0-b282-824968e884bb-kube-api-access-sgcfp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7q45z\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.658717 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7q45z\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.658774 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgcfp\" (UniqueName: \"kubernetes.io/projected/2b3c8271-0d1c-4da0-b282-824968e884bb-kube-api-access-sgcfp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7q45z\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.658840 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7q45z\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.664449 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7q45z\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.665500 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7q45z\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.680118 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgcfp\" (UniqueName: \"kubernetes.io/projected/2b3c8271-0d1c-4da0-b282-824968e884bb-kube-api-access-sgcfp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7q45z\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:14 crc kubenswrapper[4719]: I1215 12:50:14.790171 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:50:15 crc kubenswrapper[4719]: I1215 12:50:15.380511 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z"] Dec 15 12:50:16 crc kubenswrapper[4719]: I1215 12:50:16.371097 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" event={"ID":"2b3c8271-0d1c-4da0-b282-824968e884bb","Type":"ContainerStarted","Data":"4b945c0fd6a7287b76e9562a9f23350f0045adeaa034e2e1fecb3823ec91d2e6"} Dec 15 12:50:16 crc kubenswrapper[4719]: I1215 12:50:16.371393 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" event={"ID":"2b3c8271-0d1c-4da0-b282-824968e884bb","Type":"ContainerStarted","Data":"cc10ea36036d3d31e6f17dc940b0461d1aa9f4d9a709f14f64cdd34d2a3d6f08"} Dec 15 12:50:16 crc kubenswrapper[4719]: I1215 12:50:16.399651 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" podStartSLOduration=2.005383743 podStartE2EDuration="2.399635353s" podCreationTimestamp="2025-12-15 12:50:14 +0000 UTC" firstStartedPulling="2025-12-15 12:50:15.39241461 +0000 UTC m=+1976.334707650" lastFinishedPulling="2025-12-15 12:50:15.78666622 +0000 UTC m=+1976.728959260" observedRunningTime="2025-12-15 12:50:16.397878368 +0000 UTC m=+1977.340171408" watchObservedRunningTime="2025-12-15 12:50:16.399635353 +0000 UTC m=+1977.341928383" Dec 15 12:50:33 crc kubenswrapper[4719]: I1215 12:50:33.615145 4719 scope.go:117] "RemoveContainer" containerID="f08747ef15487cdbde3a0a58138ca5bc9934fbcd5aace0cff6a4e4512a4f8e46" Dec 15 12:50:33 crc kubenswrapper[4719]: I1215 12:50:33.650011 4719 scope.go:117] "RemoveContainer" containerID="3ce57c3590ea2f21308b981409e71b532cb255058381148ed40ad0853227af38" Dec 15 12:50:33 crc kubenswrapper[4719]: I1215 12:50:33.728967 4719 scope.go:117] "RemoveContainer" containerID="9054fde60087e4364449be7dabef9313151a2c1a3c452049aba7f8f452554efd" Dec 15 12:50:33 crc kubenswrapper[4719]: I1215 12:50:33.777137 4719 scope.go:117] "RemoveContainer" containerID="9da848555b8101c4361795c4e608af88e63cc10e75dbbfc98af7be184d1ed1f9" Dec 15 12:50:33 crc kubenswrapper[4719]: I1215 12:50:33.824797 4719 scope.go:117] "RemoveContainer" containerID="71ab1ddbafb71d195095c6f6fe0f9e84bfb738bddf836284b678eb37ecb0effb" Dec 15 12:50:33 crc kubenswrapper[4719]: I1215 12:50:33.892488 4719 scope.go:117] "RemoveContainer" containerID="47968e61c716203e7cb65307363c747435b970d7135f2b99b11c0f24f0411100" Dec 15 12:50:44 crc kubenswrapper[4719]: I1215 12:50:44.067189 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x6vc4"] Dec 15 12:50:44 crc kubenswrapper[4719]: I1215 12:50:44.082126 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x6vc4"] Dec 15 12:50:45 crc kubenswrapper[4719]: I1215 12:50:45.489481 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1f092f5-6e7a-408d-8985-0a900417af4e" path="/var/lib/kubelet/pods/f1f092f5-6e7a-408d-8985-0a900417af4e/volumes" Dec 15 12:50:51 crc kubenswrapper[4719]: I1215 12:50:51.357449 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:50:51 crc kubenswrapper[4719]: I1215 12:50:51.358087 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:51:11 crc kubenswrapper[4719]: I1215 12:51:11.056767 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vzsnw"] Dec 15 12:51:11 crc kubenswrapper[4719]: I1215 12:51:11.067071 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vzsnw"] Dec 15 12:51:11 crc kubenswrapper[4719]: I1215 12:51:11.505655 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30929e15-1877-4485-9366-6396d32df2ae" path="/var/lib/kubelet/pods/30929e15-1877-4485-9366-6396d32df2ae/volumes" Dec 15 12:51:18 crc kubenswrapper[4719]: I1215 12:51:18.045351 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5p65f"] Dec 15 12:51:18 crc kubenswrapper[4719]: I1215 12:51:18.060186 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5p65f"] Dec 15 12:51:19 crc kubenswrapper[4719]: I1215 12:51:19.492574 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c9e9ff2-3aa8-4640-979a-be62e16eca47" path="/var/lib/kubelet/pods/1c9e9ff2-3aa8-4640-979a-be62e16eca47/volumes" Dec 15 12:51:21 crc kubenswrapper[4719]: I1215 12:51:21.356690 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:51:21 crc kubenswrapper[4719]: I1215 12:51:21.357088 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.550651 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f6p8z"] Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.553899 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.580984 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f6p8z"] Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.699776 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-catalog-content\") pod \"redhat-operators-f6p8z\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.699850 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xztw9\" (UniqueName: \"kubernetes.io/projected/d2094f44-481e-406b-ad76-6b00c9c9aba6-kube-api-access-xztw9\") pod \"redhat-operators-f6p8z\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.700356 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-utilities\") pod \"redhat-operators-f6p8z\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.802324 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-catalog-content\") pod \"redhat-operators-f6p8z\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.802983 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-catalog-content\") pod \"redhat-operators-f6p8z\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.803168 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xztw9\" (UniqueName: \"kubernetes.io/projected/d2094f44-481e-406b-ad76-6b00c9c9aba6-kube-api-access-xztw9\") pod \"redhat-operators-f6p8z\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.803391 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-utilities\") pod \"redhat-operators-f6p8z\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.803786 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-utilities\") pod \"redhat-operators-f6p8z\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.819726 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xztw9\" (UniqueName: \"kubernetes.io/projected/d2094f44-481e-406b-ad76-6b00c9c9aba6-kube-api-access-xztw9\") pod \"redhat-operators-f6p8z\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:22 crc kubenswrapper[4719]: I1215 12:51:22.876879 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:23 crc kubenswrapper[4719]: I1215 12:51:23.327383 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f6p8z"] Dec 15 12:51:24 crc kubenswrapper[4719]: I1215 12:51:24.025204 4719 generic.go:334] "Generic (PLEG): container finished" podID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerID="efafb18f6091b7b72e3557d0df6b74683a17813041fbf479373ee39858a9ab28" exitCode=0 Dec 15 12:51:24 crc kubenswrapper[4719]: I1215 12:51:24.025383 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f6p8z" event={"ID":"d2094f44-481e-406b-ad76-6b00c9c9aba6","Type":"ContainerDied","Data":"efafb18f6091b7b72e3557d0df6b74683a17813041fbf479373ee39858a9ab28"} Dec 15 12:51:24 crc kubenswrapper[4719]: I1215 12:51:24.025543 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f6p8z" event={"ID":"d2094f44-481e-406b-ad76-6b00c9c9aba6","Type":"ContainerStarted","Data":"f1c47751478f8db386f207a43857d06ebe3a365b5a56ffd268c5580606cfa5b3"} Dec 15 12:51:25 crc kubenswrapper[4719]: I1215 12:51:25.034841 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f6p8z" event={"ID":"d2094f44-481e-406b-ad76-6b00c9c9aba6","Type":"ContainerStarted","Data":"fa9d7b020bc26fcb67bf739637633e264305dbe2e1f1c9cc37d6722afa34bb27"} Dec 15 12:51:29 crc kubenswrapper[4719]: I1215 12:51:29.065236 4719 generic.go:334] "Generic (PLEG): container finished" podID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerID="fa9d7b020bc26fcb67bf739637633e264305dbe2e1f1c9cc37d6722afa34bb27" exitCode=0 Dec 15 12:51:29 crc kubenswrapper[4719]: I1215 12:51:29.065450 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f6p8z" event={"ID":"d2094f44-481e-406b-ad76-6b00c9c9aba6","Type":"ContainerDied","Data":"fa9d7b020bc26fcb67bf739637633e264305dbe2e1f1c9cc37d6722afa34bb27"} Dec 15 12:51:30 crc kubenswrapper[4719]: I1215 12:51:30.075619 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f6p8z" event={"ID":"d2094f44-481e-406b-ad76-6b00c9c9aba6","Type":"ContainerStarted","Data":"e52b942be8ea5194fbec210e01609d8f296bbee48017b53996f9964334bd633a"} Dec 15 12:51:30 crc kubenswrapper[4719]: I1215 12:51:30.096059 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f6p8z" podStartSLOduration=2.477647026 podStartE2EDuration="8.09604578s" podCreationTimestamp="2025-12-15 12:51:22 +0000 UTC" firstStartedPulling="2025-12-15 12:51:24.030069681 +0000 UTC m=+2044.972362711" lastFinishedPulling="2025-12-15 12:51:29.648468435 +0000 UTC m=+2050.590761465" observedRunningTime="2025-12-15 12:51:30.093146289 +0000 UTC m=+2051.035439319" watchObservedRunningTime="2025-12-15 12:51:30.09604578 +0000 UTC m=+2051.038338810" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.620643 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g77x9"] Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.632705 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.655301 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g77x9"] Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.684970 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-utilities\") pod \"community-operators-g77x9\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.685151 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-792b6\" (UniqueName: \"kubernetes.io/projected/f3b44e02-bcda-4387-958b-4df9fa91c0f7-kube-api-access-792b6\") pod \"community-operators-g77x9\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.685208 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-catalog-content\") pod \"community-operators-g77x9\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.787271 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-792b6\" (UniqueName: \"kubernetes.io/projected/f3b44e02-bcda-4387-958b-4df9fa91c0f7-kube-api-access-792b6\") pod \"community-operators-g77x9\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.787346 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-catalog-content\") pod \"community-operators-g77x9\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.787407 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-utilities\") pod \"community-operators-g77x9\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.787770 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-catalog-content\") pod \"community-operators-g77x9\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.787795 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-utilities\") pod \"community-operators-g77x9\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.814029 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-792b6\" (UniqueName: \"kubernetes.io/projected/f3b44e02-bcda-4387-958b-4df9fa91c0f7-kube-api-access-792b6\") pod \"community-operators-g77x9\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:31 crc kubenswrapper[4719]: I1215 12:51:31.959928 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:32 crc kubenswrapper[4719]: I1215 12:51:32.406551 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g77x9"] Dec 15 12:51:32 crc kubenswrapper[4719]: I1215 12:51:32.877246 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:32 crc kubenswrapper[4719]: I1215 12:51:32.878512 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:33 crc kubenswrapper[4719]: I1215 12:51:33.194842 4719 generic.go:334] "Generic (PLEG): container finished" podID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerID="fcb34176c6ecc34af54b83018e02a29bc1c92cbacee137068c35dd4f785d43dd" exitCode=0 Dec 15 12:51:33 crc kubenswrapper[4719]: I1215 12:51:33.195020 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g77x9" event={"ID":"f3b44e02-bcda-4387-958b-4df9fa91c0f7","Type":"ContainerDied","Data":"fcb34176c6ecc34af54b83018e02a29bc1c92cbacee137068c35dd4f785d43dd"} Dec 15 12:51:33 crc kubenswrapper[4719]: I1215 12:51:33.195070 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g77x9" event={"ID":"f3b44e02-bcda-4387-958b-4df9fa91c0f7","Type":"ContainerStarted","Data":"f2f44f5ebce08a543a8ed7eb860a0fc6ae1ca8bcb10cd19a0502fa68c131ce1e"} Dec 15 12:51:33 crc kubenswrapper[4719]: I1215 12:51:33.921453 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f6p8z" podUID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerName="registry-server" probeResult="failure" output=< Dec 15 12:51:33 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 12:51:33 crc kubenswrapper[4719]: > Dec 15 12:51:34 crc kubenswrapper[4719]: I1215 12:51:34.043500 4719 scope.go:117] "RemoveContainer" containerID="4504d72e3231d8806da3f3047b136c2e704fefb6a67479e700f38fc11bc789df" Dec 15 12:51:34 crc kubenswrapper[4719]: I1215 12:51:34.127755 4719 scope.go:117] "RemoveContainer" containerID="166b49effe874c8c992db325f19173be0def7b459985ad660ca508d4805b4f8f" Dec 15 12:51:34 crc kubenswrapper[4719]: I1215 12:51:34.204469 4719 scope.go:117] "RemoveContainer" containerID="395f7c82e3a73edfeb1672d4ba7a438f9553821b569be491822069549733d9c0" Dec 15 12:51:35 crc kubenswrapper[4719]: I1215 12:51:35.224643 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g77x9" event={"ID":"f3b44e02-bcda-4387-958b-4df9fa91c0f7","Type":"ContainerStarted","Data":"9b0235c84a8b2d9cecbcc960017fe094750ce678c3b4facf733cdc2ae355c969"} Dec 15 12:51:36 crc kubenswrapper[4719]: I1215 12:51:36.233770 4719 generic.go:334] "Generic (PLEG): container finished" podID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerID="9b0235c84a8b2d9cecbcc960017fe094750ce678c3b4facf733cdc2ae355c969" exitCode=0 Dec 15 12:51:36 crc kubenswrapper[4719]: I1215 12:51:36.233823 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g77x9" event={"ID":"f3b44e02-bcda-4387-958b-4df9fa91c0f7","Type":"ContainerDied","Data":"9b0235c84a8b2d9cecbcc960017fe094750ce678c3b4facf733cdc2ae355c969"} Dec 15 12:51:37 crc kubenswrapper[4719]: I1215 12:51:37.243779 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g77x9" event={"ID":"f3b44e02-bcda-4387-958b-4df9fa91c0f7","Type":"ContainerStarted","Data":"d7ab779eb0c04b66195473c3ac28be5e23577b90a47ff160a48774d885aeb8a9"} Dec 15 12:51:37 crc kubenswrapper[4719]: I1215 12:51:37.270670 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g77x9" podStartSLOduration=2.742404651 podStartE2EDuration="6.27065337s" podCreationTimestamp="2025-12-15 12:51:31 +0000 UTC" firstStartedPulling="2025-12-15 12:51:33.197125585 +0000 UTC m=+2054.139418615" lastFinishedPulling="2025-12-15 12:51:36.725374294 +0000 UTC m=+2057.667667334" observedRunningTime="2025-12-15 12:51:37.262003478 +0000 UTC m=+2058.204296508" watchObservedRunningTime="2025-12-15 12:51:37.27065337 +0000 UTC m=+2058.212946400" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.283271 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s4wnf"] Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.288411 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.296444 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s4wnf"] Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.336116 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvwxw\" (UniqueName: \"kubernetes.io/projected/410383a1-1026-4f05-9287-f6e678dc5666-kube-api-access-jvwxw\") pod \"certified-operators-s4wnf\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.336247 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-catalog-content\") pod \"certified-operators-s4wnf\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.336393 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-utilities\") pod \"certified-operators-s4wnf\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.439532 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-catalog-content\") pod \"certified-operators-s4wnf\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.439627 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-utilities\") pod \"certified-operators-s4wnf\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.439694 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvwxw\" (UniqueName: \"kubernetes.io/projected/410383a1-1026-4f05-9287-f6e678dc5666-kube-api-access-jvwxw\") pod \"certified-operators-s4wnf\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.440351 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-catalog-content\") pod \"certified-operators-s4wnf\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.440461 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-utilities\") pod \"certified-operators-s4wnf\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.480829 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvwxw\" (UniqueName: \"kubernetes.io/projected/410383a1-1026-4f05-9287-f6e678dc5666-kube-api-access-jvwxw\") pod \"certified-operators-s4wnf\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:39 crc kubenswrapper[4719]: I1215 12:51:39.611012 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:40 crc kubenswrapper[4719]: I1215 12:51:40.286945 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s4wnf"] Dec 15 12:51:41 crc kubenswrapper[4719]: I1215 12:51:41.274821 4719 generic.go:334] "Generic (PLEG): container finished" podID="410383a1-1026-4f05-9287-f6e678dc5666" containerID="1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228" exitCode=0 Dec 15 12:51:41 crc kubenswrapper[4719]: I1215 12:51:41.274944 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4wnf" event={"ID":"410383a1-1026-4f05-9287-f6e678dc5666","Type":"ContainerDied","Data":"1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228"} Dec 15 12:51:41 crc kubenswrapper[4719]: I1215 12:51:41.275360 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4wnf" event={"ID":"410383a1-1026-4f05-9287-f6e678dc5666","Type":"ContainerStarted","Data":"70d5a48cea624eab5df0b34bac0a71ebda3790aab217c721899e01578bd29b86"} Dec 15 12:51:41 crc kubenswrapper[4719]: I1215 12:51:41.960400 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:41 crc kubenswrapper[4719]: I1215 12:51:41.960492 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:42 crc kubenswrapper[4719]: I1215 12:51:42.024576 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:42 crc kubenswrapper[4719]: I1215 12:51:42.290285 4719 generic.go:334] "Generic (PLEG): container finished" podID="2b3c8271-0d1c-4da0-b282-824968e884bb" containerID="4b945c0fd6a7287b76e9562a9f23350f0045adeaa034e2e1fecb3823ec91d2e6" exitCode=0 Dec 15 12:51:42 crc kubenswrapper[4719]: I1215 12:51:42.290367 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" event={"ID":"2b3c8271-0d1c-4da0-b282-824968e884bb","Type":"ContainerDied","Data":"4b945c0fd6a7287b76e9562a9f23350f0045adeaa034e2e1fecb3823ec91d2e6"} Dec 15 12:51:42 crc kubenswrapper[4719]: I1215 12:51:42.346499 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:42 crc kubenswrapper[4719]: I1215 12:51:42.927788 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:42 crc kubenswrapper[4719]: I1215 12:51:42.980175 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:43 crc kubenswrapper[4719]: I1215 12:51:43.303968 4719 generic.go:334] "Generic (PLEG): container finished" podID="410383a1-1026-4f05-9287-f6e678dc5666" containerID="635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01" exitCode=0 Dec 15 12:51:43 crc kubenswrapper[4719]: I1215 12:51:43.304081 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4wnf" event={"ID":"410383a1-1026-4f05-9287-f6e678dc5666","Type":"ContainerDied","Data":"635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01"} Dec 15 12:51:43 crc kubenswrapper[4719]: I1215 12:51:43.804671 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:51:43 crc kubenswrapper[4719]: I1215 12:51:43.977131 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-inventory\") pod \"2b3c8271-0d1c-4da0-b282-824968e884bb\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " Dec 15 12:51:43 crc kubenswrapper[4719]: I1215 12:51:43.977452 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgcfp\" (UniqueName: \"kubernetes.io/projected/2b3c8271-0d1c-4da0-b282-824968e884bb-kube-api-access-sgcfp\") pod \"2b3c8271-0d1c-4da0-b282-824968e884bb\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " Dec 15 12:51:43 crc kubenswrapper[4719]: I1215 12:51:43.977679 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-ssh-key\") pod \"2b3c8271-0d1c-4da0-b282-824968e884bb\" (UID: \"2b3c8271-0d1c-4da0-b282-824968e884bb\") " Dec 15 12:51:43 crc kubenswrapper[4719]: I1215 12:51:43.982921 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b3c8271-0d1c-4da0-b282-824968e884bb-kube-api-access-sgcfp" (OuterVolumeSpecName: "kube-api-access-sgcfp") pod "2b3c8271-0d1c-4da0-b282-824968e884bb" (UID: "2b3c8271-0d1c-4da0-b282-824968e884bb"). InnerVolumeSpecName "kube-api-access-sgcfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.003303 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-inventory" (OuterVolumeSpecName: "inventory") pod "2b3c8271-0d1c-4da0-b282-824968e884bb" (UID: "2b3c8271-0d1c-4da0-b282-824968e884bb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.004407 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2b3c8271-0d1c-4da0-b282-824968e884bb" (UID: "2b3c8271-0d1c-4da0-b282-824968e884bb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.079773 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.079809 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b3c8271-0d1c-4da0-b282-824968e884bb-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.079824 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgcfp\" (UniqueName: \"kubernetes.io/projected/2b3c8271-0d1c-4da0-b282-824968e884bb-kube-api-access-sgcfp\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.317562 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.317654 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7q45z" event={"ID":"2b3c8271-0d1c-4da0-b282-824968e884bb","Type":"ContainerDied","Data":"cc10ea36036d3d31e6f17dc940b0461d1aa9f4d9a709f14f64cdd34d2a3d6f08"} Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.317705 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc10ea36036d3d31e6f17dc940b0461d1aa9f4d9a709f14f64cdd34d2a3d6f08" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.356678 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g77x9"] Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.449574 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl"] Dec 15 12:51:44 crc kubenswrapper[4719]: E1215 12:51:44.450002 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b3c8271-0d1c-4da0-b282-824968e884bb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.450018 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3c8271-0d1c-4da0-b282-824968e884bb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.450197 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b3c8271-0d1c-4da0-b282-824968e884bb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.450958 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.453562 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.453569 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.453582 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.454778 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.461401 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl"] Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.589746 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mddww\" (UniqueName: \"kubernetes.io/projected/2c709af4-4a8c-4cb9-955e-37dfe843569b-kube-api-access-mddww\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.589914 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.589968 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.691642 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mddww\" (UniqueName: \"kubernetes.io/projected/2c709af4-4a8c-4cb9-955e-37dfe843569b-kube-api-access-mddww\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.691766 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.691812 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.697227 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.703387 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.713956 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mddww\" (UniqueName: \"kubernetes.io/projected/2c709af4-4a8c-4cb9-955e-37dfe843569b-kube-api-access-mddww\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:44 crc kubenswrapper[4719]: I1215 12:51:44.789771 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:45 crc kubenswrapper[4719]: I1215 12:51:45.328058 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g77x9" podUID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerName="registry-server" containerID="cri-o://d7ab779eb0c04b66195473c3ac28be5e23577b90a47ff160a48774d885aeb8a9" gracePeriod=2 Dec 15 12:51:45 crc kubenswrapper[4719]: I1215 12:51:45.631257 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl"] Dec 15 12:51:45 crc kubenswrapper[4719]: W1215 12:51:45.634848 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c709af4_4a8c_4cb9_955e_37dfe843569b.slice/crio-5307059edd00c3f86895d737cb913daf0c16372194cd37e4bbfb8ef92b91d9b6 WatchSource:0}: Error finding container 5307059edd00c3f86895d737cb913daf0c16372194cd37e4bbfb8ef92b91d9b6: Status 404 returned error can't find the container with id 5307059edd00c3f86895d737cb913daf0c16372194cd37e4bbfb8ef92b91d9b6 Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.156076 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f6p8z"] Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.156630 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f6p8z" podUID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerName="registry-server" containerID="cri-o://e52b942be8ea5194fbec210e01609d8f296bbee48017b53996f9964334bd633a" gracePeriod=2 Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.336815 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" event={"ID":"2c709af4-4a8c-4cb9-955e-37dfe843569b","Type":"ContainerStarted","Data":"5307059edd00c3f86895d737cb913daf0c16372194cd37e4bbfb8ef92b91d9b6"} Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.339178 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4wnf" event={"ID":"410383a1-1026-4f05-9287-f6e678dc5666","Type":"ContainerStarted","Data":"6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc"} Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.342315 4719 generic.go:334] "Generic (PLEG): container finished" podID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerID="e52b942be8ea5194fbec210e01609d8f296bbee48017b53996f9964334bd633a" exitCode=0 Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.342390 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f6p8z" event={"ID":"d2094f44-481e-406b-ad76-6b00c9c9aba6","Type":"ContainerDied","Data":"e52b942be8ea5194fbec210e01609d8f296bbee48017b53996f9964334bd633a"} Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.345144 4719 generic.go:334] "Generic (PLEG): container finished" podID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerID="d7ab779eb0c04b66195473c3ac28be5e23577b90a47ff160a48774d885aeb8a9" exitCode=0 Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.345207 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g77x9" event={"ID":"f3b44e02-bcda-4387-958b-4df9fa91c0f7","Type":"ContainerDied","Data":"d7ab779eb0c04b66195473c3ac28be5e23577b90a47ff160a48774d885aeb8a9"} Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.345236 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g77x9" event={"ID":"f3b44e02-bcda-4387-958b-4df9fa91c0f7","Type":"ContainerDied","Data":"f2f44f5ebce08a543a8ed7eb860a0fc6ae1ca8bcb10cd19a0502fa68c131ce1e"} Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.345250 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2f44f5ebce08a543a8ed7eb860a0fc6ae1ca8bcb10cd19a0502fa68c131ce1e" Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.362094 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s4wnf" podStartSLOduration=3.531552279 podStartE2EDuration="7.362078885s" podCreationTimestamp="2025-12-15 12:51:39 +0000 UTC" firstStartedPulling="2025-12-15 12:51:41.276972013 +0000 UTC m=+2062.219265043" lastFinishedPulling="2025-12-15 12:51:45.107498619 +0000 UTC m=+2066.049791649" observedRunningTime="2025-12-15 12:51:46.360268618 +0000 UTC m=+2067.302561648" watchObservedRunningTime="2025-12-15 12:51:46.362078885 +0000 UTC m=+2067.304371915" Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.393753 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.528718 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-792b6\" (UniqueName: \"kubernetes.io/projected/f3b44e02-bcda-4387-958b-4df9fa91c0f7-kube-api-access-792b6\") pod \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.529158 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-catalog-content\") pod \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.529289 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-utilities\") pod \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\" (UID: \"f3b44e02-bcda-4387-958b-4df9fa91c0f7\") " Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.530701 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-utilities" (OuterVolumeSpecName: "utilities") pod "f3b44e02-bcda-4387-958b-4df9fa91c0f7" (UID: "f3b44e02-bcda-4387-958b-4df9fa91c0f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.535070 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3b44e02-bcda-4387-958b-4df9fa91c0f7-kube-api-access-792b6" (OuterVolumeSpecName: "kube-api-access-792b6") pod "f3b44e02-bcda-4387-958b-4df9fa91c0f7" (UID: "f3b44e02-bcda-4387-958b-4df9fa91c0f7"). InnerVolumeSpecName "kube-api-access-792b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.600131 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3b44e02-bcda-4387-958b-4df9fa91c0f7" (UID: "f3b44e02-bcda-4387-958b-4df9fa91c0f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.646387 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.646648 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3b44e02-bcda-4387-958b-4df9fa91c0f7-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:46 crc kubenswrapper[4719]: I1215 12:51:46.646713 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-792b6\" (UniqueName: \"kubernetes.io/projected/f3b44e02-bcda-4387-958b-4df9fa91c0f7-kube-api-access-792b6\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.106463 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.292608 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-catalog-content\") pod \"d2094f44-481e-406b-ad76-6b00c9c9aba6\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.293051 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xztw9\" (UniqueName: \"kubernetes.io/projected/d2094f44-481e-406b-ad76-6b00c9c9aba6-kube-api-access-xztw9\") pod \"d2094f44-481e-406b-ad76-6b00c9c9aba6\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.293090 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-utilities\") pod \"d2094f44-481e-406b-ad76-6b00c9c9aba6\" (UID: \"d2094f44-481e-406b-ad76-6b00c9c9aba6\") " Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.293501 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-utilities" (OuterVolumeSpecName: "utilities") pod "d2094f44-481e-406b-ad76-6b00c9c9aba6" (UID: "d2094f44-481e-406b-ad76-6b00c9c9aba6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.302710 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2094f44-481e-406b-ad76-6b00c9c9aba6-kube-api-access-xztw9" (OuterVolumeSpecName: "kube-api-access-xztw9") pod "d2094f44-481e-406b-ad76-6b00c9c9aba6" (UID: "d2094f44-481e-406b-ad76-6b00c9c9aba6"). InnerVolumeSpecName "kube-api-access-xztw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.355255 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g77x9" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.361092 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f6p8z" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.360967 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f6p8z" event={"ID":"d2094f44-481e-406b-ad76-6b00c9c9aba6","Type":"ContainerDied","Data":"f1c47751478f8db386f207a43857d06ebe3a365b5a56ffd268c5580606cfa5b3"} Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.363000 4719 scope.go:117] "RemoveContainer" containerID="e52b942be8ea5194fbec210e01609d8f296bbee48017b53996f9964334bd633a" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.395397 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.395436 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xztw9\" (UniqueName: \"kubernetes.io/projected/d2094f44-481e-406b-ad76-6b00c9c9aba6-kube-api-access-xztw9\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.396532 4719 scope.go:117] "RemoveContainer" containerID="fa9d7b020bc26fcb67bf739637633e264305dbe2e1f1c9cc37d6722afa34bb27" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.396746 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g77x9"] Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.404414 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g77x9"] Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.414958 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2094f44-481e-406b-ad76-6b00c9c9aba6" (UID: "d2094f44-481e-406b-ad76-6b00c9c9aba6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.425180 4719 scope.go:117] "RemoveContainer" containerID="efafb18f6091b7b72e3557d0df6b74683a17813041fbf479373ee39858a9ab28" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.488975 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" path="/var/lib/kubelet/pods/f3b44e02-bcda-4387-958b-4df9fa91c0f7/volumes" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.497718 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2094f44-481e-406b-ad76-6b00c9c9aba6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.687145 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f6p8z"] Dec 15 12:51:47 crc kubenswrapper[4719]: I1215 12:51:47.696295 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f6p8z"] Dec 15 12:51:49 crc kubenswrapper[4719]: I1215 12:51:49.376127 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" event={"ID":"2c709af4-4a8c-4cb9-955e-37dfe843569b","Type":"ContainerStarted","Data":"12ce93e56c25ee983a28c5655185530ad3a502457e12457fabfbc9f3c652fffb"} Dec 15 12:51:49 crc kubenswrapper[4719]: I1215 12:51:49.402439 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" podStartSLOduration=2.689316208 podStartE2EDuration="5.402421071s" podCreationTimestamp="2025-12-15 12:51:44 +0000 UTC" firstStartedPulling="2025-12-15 12:51:45.638458764 +0000 UTC m=+2066.580751784" lastFinishedPulling="2025-12-15 12:51:48.351563597 +0000 UTC m=+2069.293856647" observedRunningTime="2025-12-15 12:51:49.395877475 +0000 UTC m=+2070.338170505" watchObservedRunningTime="2025-12-15 12:51:49.402421071 +0000 UTC m=+2070.344714091" Dec 15 12:51:49 crc kubenswrapper[4719]: I1215 12:51:49.485750 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2094f44-481e-406b-ad76-6b00c9c9aba6" path="/var/lib/kubelet/pods/d2094f44-481e-406b-ad76-6b00c9c9aba6/volumes" Dec 15 12:51:49 crc kubenswrapper[4719]: I1215 12:51:49.612567 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:49 crc kubenswrapper[4719]: I1215 12:51:49.612689 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:49 crc kubenswrapper[4719]: I1215 12:51:49.661700 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:50 crc kubenswrapper[4719]: I1215 12:51:50.433738 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:51 crc kubenswrapper[4719]: I1215 12:51:51.357237 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:51:51 crc kubenswrapper[4719]: I1215 12:51:51.357294 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:51:51 crc kubenswrapper[4719]: I1215 12:51:51.357344 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:51:51 crc kubenswrapper[4719]: I1215 12:51:51.358126 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"23da24f9e77dd91d08a2fcb3b8667d259f3c0a3ecdb0953729476936101d7499"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 12:51:51 crc kubenswrapper[4719]: I1215 12:51:51.358196 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://23da24f9e77dd91d08a2fcb3b8667d259f3c0a3ecdb0953729476936101d7499" gracePeriod=600 Dec 15 12:51:52 crc kubenswrapper[4719]: I1215 12:51:52.403161 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="23da24f9e77dd91d08a2fcb3b8667d259f3c0a3ecdb0953729476936101d7499" exitCode=0 Dec 15 12:51:52 crc kubenswrapper[4719]: I1215 12:51:52.403275 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"23da24f9e77dd91d08a2fcb3b8667d259f3c0a3ecdb0953729476936101d7499"} Dec 15 12:51:52 crc kubenswrapper[4719]: I1215 12:51:52.403711 4719 scope.go:117] "RemoveContainer" containerID="038a070537c37f3976255ec1b60b1271cf7c622627d1b6b832554d4f25ec337b" Dec 15 12:51:52 crc kubenswrapper[4719]: I1215 12:51:52.549672 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s4wnf"] Dec 15 12:51:53 crc kubenswrapper[4719]: I1215 12:51:53.423213 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s4wnf" podUID="410383a1-1026-4f05-9287-f6e678dc5666" containerName="registry-server" containerID="cri-o://6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc" gracePeriod=2 Dec 15 12:51:53 crc kubenswrapper[4719]: I1215 12:51:53.423904 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58"} Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.337666 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.434269 4719 generic.go:334] "Generic (PLEG): container finished" podID="410383a1-1026-4f05-9287-f6e678dc5666" containerID="6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc" exitCode=0 Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.434363 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s4wnf" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.434781 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4wnf" event={"ID":"410383a1-1026-4f05-9287-f6e678dc5666","Type":"ContainerDied","Data":"6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc"} Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.434805 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s4wnf" event={"ID":"410383a1-1026-4f05-9287-f6e678dc5666","Type":"ContainerDied","Data":"70d5a48cea624eab5df0b34bac0a71ebda3790aab217c721899e01578bd29b86"} Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.434821 4719 scope.go:117] "RemoveContainer" containerID="6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.440777 4719 generic.go:334] "Generic (PLEG): container finished" podID="2c709af4-4a8c-4cb9-955e-37dfe843569b" containerID="12ce93e56c25ee983a28c5655185530ad3a502457e12457fabfbc9f3c652fffb" exitCode=0 Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.440882 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" event={"ID":"2c709af4-4a8c-4cb9-955e-37dfe843569b","Type":"ContainerDied","Data":"12ce93e56c25ee983a28c5655185530ad3a502457e12457fabfbc9f3c652fffb"} Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.483156 4719 scope.go:117] "RemoveContainer" containerID="635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.503403 4719 scope.go:117] "RemoveContainer" containerID="1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.532690 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvwxw\" (UniqueName: \"kubernetes.io/projected/410383a1-1026-4f05-9287-f6e678dc5666-kube-api-access-jvwxw\") pod \"410383a1-1026-4f05-9287-f6e678dc5666\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.533065 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-catalog-content\") pod \"410383a1-1026-4f05-9287-f6e678dc5666\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.533122 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-utilities\") pod \"410383a1-1026-4f05-9287-f6e678dc5666\" (UID: \"410383a1-1026-4f05-9287-f6e678dc5666\") " Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.533940 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-utilities" (OuterVolumeSpecName: "utilities") pod "410383a1-1026-4f05-9287-f6e678dc5666" (UID: "410383a1-1026-4f05-9287-f6e678dc5666"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.539850 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/410383a1-1026-4f05-9287-f6e678dc5666-kube-api-access-jvwxw" (OuterVolumeSpecName: "kube-api-access-jvwxw") pod "410383a1-1026-4f05-9287-f6e678dc5666" (UID: "410383a1-1026-4f05-9287-f6e678dc5666"). InnerVolumeSpecName "kube-api-access-jvwxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.587419 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "410383a1-1026-4f05-9287-f6e678dc5666" (UID: "410383a1-1026-4f05-9287-f6e678dc5666"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.614870 4719 scope.go:117] "RemoveContainer" containerID="6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc" Dec 15 12:51:54 crc kubenswrapper[4719]: E1215 12:51:54.615365 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc\": container with ID starting with 6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc not found: ID does not exist" containerID="6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.615415 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc"} err="failed to get container status \"6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc\": rpc error: code = NotFound desc = could not find container \"6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc\": container with ID starting with 6057ea3d45243f0c4fa553fb078108c5414b68b7a72b4a0c80789330ea90d3dc not found: ID does not exist" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.615443 4719 scope.go:117] "RemoveContainer" containerID="635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01" Dec 15 12:51:54 crc kubenswrapper[4719]: E1215 12:51:54.615880 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01\": container with ID starting with 635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01 not found: ID does not exist" containerID="635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.615933 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01"} err="failed to get container status \"635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01\": rpc error: code = NotFound desc = could not find container \"635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01\": container with ID starting with 635200965fb5c23c86a5d679f8984b46aef517eda24e193609eeabbf6a7c5b01 not found: ID does not exist" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.615950 4719 scope.go:117] "RemoveContainer" containerID="1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228" Dec 15 12:51:54 crc kubenswrapper[4719]: E1215 12:51:54.616415 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228\": container with ID starting with 1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228 not found: ID does not exist" containerID="1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.616443 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228"} err="failed to get container status \"1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228\": rpc error: code = NotFound desc = could not find container \"1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228\": container with ID starting with 1d3f84db47c3ecd923195856eb83cb848737f03d0d63301deb44935f798c9228 not found: ID does not exist" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.635732 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvwxw\" (UniqueName: \"kubernetes.io/projected/410383a1-1026-4f05-9287-f6e678dc5666-kube-api-access-jvwxw\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.635804 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.635817 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410383a1-1026-4f05-9287-f6e678dc5666-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.765959 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s4wnf"] Dec 15 12:51:54 crc kubenswrapper[4719]: I1215 12:51:54.772818 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s4wnf"] Dec 15 12:51:55 crc kubenswrapper[4719]: I1215 12:51:55.490269 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="410383a1-1026-4f05-9287-f6e678dc5666" path="/var/lib/kubelet/pods/410383a1-1026-4f05-9287-f6e678dc5666/volumes" Dec 15 12:51:55 crc kubenswrapper[4719]: I1215 12:51:55.878049 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.064191 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-inventory\") pod \"2c709af4-4a8c-4cb9-955e-37dfe843569b\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.064739 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-ssh-key\") pod \"2c709af4-4a8c-4cb9-955e-37dfe843569b\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.064803 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mddww\" (UniqueName: \"kubernetes.io/projected/2c709af4-4a8c-4cb9-955e-37dfe843569b-kube-api-access-mddww\") pod \"2c709af4-4a8c-4cb9-955e-37dfe843569b\" (UID: \"2c709af4-4a8c-4cb9-955e-37dfe843569b\") " Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.069957 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c709af4-4a8c-4cb9-955e-37dfe843569b-kube-api-access-mddww" (OuterVolumeSpecName: "kube-api-access-mddww") pod "2c709af4-4a8c-4cb9-955e-37dfe843569b" (UID: "2c709af4-4a8c-4cb9-955e-37dfe843569b"). InnerVolumeSpecName "kube-api-access-mddww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.090039 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-inventory" (OuterVolumeSpecName: "inventory") pod "2c709af4-4a8c-4cb9-955e-37dfe843569b" (UID: "2c709af4-4a8c-4cb9-955e-37dfe843569b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.098132 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2c709af4-4a8c-4cb9-955e-37dfe843569b" (UID: "2c709af4-4a8c-4cb9-955e-37dfe843569b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.167328 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.167357 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c709af4-4a8c-4cb9-955e-37dfe843569b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.167366 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mddww\" (UniqueName: \"kubernetes.io/projected/2c709af4-4a8c-4cb9-955e-37dfe843569b-kube-api-access-mddww\") on node \"crc\" DevicePath \"\"" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.468108 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" event={"ID":"2c709af4-4a8c-4cb9-955e-37dfe843569b","Type":"ContainerDied","Data":"5307059edd00c3f86895d737cb913daf0c16372194cd37e4bbfb8ef92b91d9b6"} Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.468165 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5307059edd00c3f86895d737cb913daf0c16372194cd37e4bbfb8ef92b91d9b6" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.468255 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.534909 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg"] Dec 15 12:51:56 crc kubenswrapper[4719]: E1215 12:51:56.535236 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerName="extract-content" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535247 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerName="extract-content" Dec 15 12:51:56 crc kubenswrapper[4719]: E1215 12:51:56.535269 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410383a1-1026-4f05-9287-f6e678dc5666" containerName="registry-server" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535276 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="410383a1-1026-4f05-9287-f6e678dc5666" containerName="registry-server" Dec 15 12:51:56 crc kubenswrapper[4719]: E1215 12:51:56.535287 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerName="registry-server" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535294 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerName="registry-server" Dec 15 12:51:56 crc kubenswrapper[4719]: E1215 12:51:56.535303 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerName="registry-server" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535308 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerName="registry-server" Dec 15 12:51:56 crc kubenswrapper[4719]: E1215 12:51:56.535316 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerName="extract-utilities" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535322 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerName="extract-utilities" Dec 15 12:51:56 crc kubenswrapper[4719]: E1215 12:51:56.535332 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c709af4-4a8c-4cb9-955e-37dfe843569b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535338 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c709af4-4a8c-4cb9-955e-37dfe843569b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 15 12:51:56 crc kubenswrapper[4719]: E1215 12:51:56.535349 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410383a1-1026-4f05-9287-f6e678dc5666" containerName="extract-utilities" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535354 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="410383a1-1026-4f05-9287-f6e678dc5666" containerName="extract-utilities" Dec 15 12:51:56 crc kubenswrapper[4719]: E1215 12:51:56.535368 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410383a1-1026-4f05-9287-f6e678dc5666" containerName="extract-content" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535374 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="410383a1-1026-4f05-9287-f6e678dc5666" containerName="extract-content" Dec 15 12:51:56 crc kubenswrapper[4719]: E1215 12:51:56.535386 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerName="extract-utilities" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535391 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerName="extract-utilities" Dec 15 12:51:56 crc kubenswrapper[4719]: E1215 12:51:56.535403 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerName="extract-content" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535410 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerName="extract-content" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535564 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3b44e02-bcda-4387-958b-4df9fa91c0f7" containerName="registry-server" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535578 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c709af4-4a8c-4cb9-955e-37dfe843569b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535596 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="410383a1-1026-4f05-9287-f6e678dc5666" containerName="registry-server" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.535609 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2094f44-481e-406b-ad76-6b00c9c9aba6" containerName="registry-server" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.536164 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.538426 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.539789 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.539983 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.543225 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.559036 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg"] Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.684534 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hsfw\" (UniqueName: \"kubernetes.io/projected/2e7c3312-b561-431c-8794-c63098dbe33d-kube-api-access-5hsfw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gv6zg\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.684637 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gv6zg\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.684687 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gv6zg\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.786388 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hsfw\" (UniqueName: \"kubernetes.io/projected/2e7c3312-b561-431c-8794-c63098dbe33d-kube-api-access-5hsfw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gv6zg\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.786440 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gv6zg\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.786463 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gv6zg\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.792307 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gv6zg\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.799127 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gv6zg\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.803662 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hsfw\" (UniqueName: \"kubernetes.io/projected/2e7c3312-b561-431c-8794-c63098dbe33d-kube-api-access-5hsfw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gv6zg\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:56 crc kubenswrapper[4719]: I1215 12:51:56.863334 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:51:57 crc kubenswrapper[4719]: I1215 12:51:57.349351 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg"] Dec 15 12:51:57 crc kubenswrapper[4719]: I1215 12:51:57.489186 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" event={"ID":"2e7c3312-b561-431c-8794-c63098dbe33d","Type":"ContainerStarted","Data":"ca720b927c48e63739e51c145b3611652ddb1b24ebb750cc92a91a26a83fe7d0"} Dec 15 12:51:58 crc kubenswrapper[4719]: I1215 12:51:58.059541 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-j5gcb"] Dec 15 12:51:58 crc kubenswrapper[4719]: I1215 12:51:58.079447 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-j5gcb"] Dec 15 12:51:58 crc kubenswrapper[4719]: I1215 12:51:58.489714 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" event={"ID":"2e7c3312-b561-431c-8794-c63098dbe33d","Type":"ContainerStarted","Data":"c1641e87f74b16bbcc5ed21b1f62bc7030f08b35f451fa8ddf85a39f194c2e3f"} Dec 15 12:51:58 crc kubenswrapper[4719]: I1215 12:51:58.556594 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" podStartSLOduration=1.9126066019999999 podStartE2EDuration="2.55657046s" podCreationTimestamp="2025-12-15 12:51:56 +0000 UTC" firstStartedPulling="2025-12-15 12:51:57.357614881 +0000 UTC m=+2078.299907911" lastFinishedPulling="2025-12-15 12:51:58.001578739 +0000 UTC m=+2078.943871769" observedRunningTime="2025-12-15 12:51:58.552356367 +0000 UTC m=+2079.494649407" watchObservedRunningTime="2025-12-15 12:51:58.55657046 +0000 UTC m=+2079.498863520" Dec 15 12:51:59 crc kubenswrapper[4719]: I1215 12:51:59.490896 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33983783-3b6b-4a9e-93c0-2cf842823bf3" path="/var/lib/kubelet/pods/33983783-3b6b-4a9e-93c0-2cf842823bf3/volumes" Dec 15 12:52:34 crc kubenswrapper[4719]: I1215 12:52:34.312021 4719 scope.go:117] "RemoveContainer" containerID="e98e660ab24653045210f1db1ceaca509ac69edeb24cfadc11978986dc3e8fc2" Dec 15 12:52:39 crc kubenswrapper[4719]: I1215 12:52:39.874832 4719 generic.go:334] "Generic (PLEG): container finished" podID="2e7c3312-b561-431c-8794-c63098dbe33d" containerID="c1641e87f74b16bbcc5ed21b1f62bc7030f08b35f451fa8ddf85a39f194c2e3f" exitCode=0 Dec 15 12:52:39 crc kubenswrapper[4719]: I1215 12:52:39.874928 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" event={"ID":"2e7c3312-b561-431c-8794-c63098dbe33d","Type":"ContainerDied","Data":"c1641e87f74b16bbcc5ed21b1f62bc7030f08b35f451fa8ddf85a39f194c2e3f"} Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.296818 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.389891 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-ssh-key\") pod \"2e7c3312-b561-431c-8794-c63098dbe33d\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.390358 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hsfw\" (UniqueName: \"kubernetes.io/projected/2e7c3312-b561-431c-8794-c63098dbe33d-kube-api-access-5hsfw\") pod \"2e7c3312-b561-431c-8794-c63098dbe33d\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.390400 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-inventory\") pod \"2e7c3312-b561-431c-8794-c63098dbe33d\" (UID: \"2e7c3312-b561-431c-8794-c63098dbe33d\") " Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.403182 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e7c3312-b561-431c-8794-c63098dbe33d-kube-api-access-5hsfw" (OuterVolumeSpecName: "kube-api-access-5hsfw") pod "2e7c3312-b561-431c-8794-c63098dbe33d" (UID: "2e7c3312-b561-431c-8794-c63098dbe33d"). InnerVolumeSpecName "kube-api-access-5hsfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.416520 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-inventory" (OuterVolumeSpecName: "inventory") pod "2e7c3312-b561-431c-8794-c63098dbe33d" (UID: "2e7c3312-b561-431c-8794-c63098dbe33d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.431445 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2e7c3312-b561-431c-8794-c63098dbe33d" (UID: "2e7c3312-b561-431c-8794-c63098dbe33d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.499090 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.499121 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e7c3312-b561-431c-8794-c63098dbe33d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.499131 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hsfw\" (UniqueName: \"kubernetes.io/projected/2e7c3312-b561-431c-8794-c63098dbe33d-kube-api-access-5hsfw\") on node \"crc\" DevicePath \"\"" Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.893902 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" event={"ID":"2e7c3312-b561-431c-8794-c63098dbe33d","Type":"ContainerDied","Data":"ca720b927c48e63739e51c145b3611652ddb1b24ebb750cc92a91a26a83fe7d0"} Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.893963 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca720b927c48e63739e51c145b3611652ddb1b24ebb750cc92a91a26a83fe7d0" Dec 15 12:52:41 crc kubenswrapper[4719]: I1215 12:52:41.894011 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gv6zg" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.007936 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj"] Dec 15 12:52:42 crc kubenswrapper[4719]: E1215 12:52:42.008376 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e7c3312-b561-431c-8794-c63098dbe33d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.008404 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e7c3312-b561-431c-8794-c63098dbe33d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.008673 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e7c3312-b561-431c-8794-c63098dbe33d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.009420 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.011790 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.012066 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.012187 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.013177 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.026969 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj"] Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.110623 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.110822 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.110898 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlnd9\" (UniqueName: \"kubernetes.io/projected/611b65c2-c554-4f37-a644-41bcbd27ad46-kube-api-access-vlnd9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.212968 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.213117 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.213152 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlnd9\" (UniqueName: \"kubernetes.io/projected/611b65c2-c554-4f37-a644-41bcbd27ad46-kube-api-access-vlnd9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.219664 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.225435 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.230026 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlnd9\" (UniqueName: \"kubernetes.io/projected/611b65c2-c554-4f37-a644-41bcbd27ad46-kube-api-access-vlnd9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.338123 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.880919 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj"] Dec 15 12:52:42 crc kubenswrapper[4719]: I1215 12:52:42.902580 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" event={"ID":"611b65c2-c554-4f37-a644-41bcbd27ad46","Type":"ContainerStarted","Data":"954a00f9f4b05fb2bb88bce4acb7b6d667f072e71060270580a9e0d6404e6f44"} Dec 15 12:52:43 crc kubenswrapper[4719]: I1215 12:52:43.911105 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" event={"ID":"611b65c2-c554-4f37-a644-41bcbd27ad46","Type":"ContainerStarted","Data":"4cf6a93d4de44f32f710d1a2d5b04831fb802e97c30b7fc7cd229bfe8afaa2a1"} Dec 15 12:52:43 crc kubenswrapper[4719]: I1215 12:52:43.939432 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" podStartSLOduration=2.422896452 podStartE2EDuration="2.939415034s" podCreationTimestamp="2025-12-15 12:52:41 +0000 UTC" firstStartedPulling="2025-12-15 12:52:42.876843912 +0000 UTC m=+2123.819136942" lastFinishedPulling="2025-12-15 12:52:43.393362494 +0000 UTC m=+2124.335655524" observedRunningTime="2025-12-15 12:52:43.933617592 +0000 UTC m=+2124.875910622" watchObservedRunningTime="2025-12-15 12:52:43.939415034 +0000 UTC m=+2124.881708064" Dec 15 12:53:40 crc kubenswrapper[4719]: I1215 12:53:40.404894 4719 generic.go:334] "Generic (PLEG): container finished" podID="611b65c2-c554-4f37-a644-41bcbd27ad46" containerID="4cf6a93d4de44f32f710d1a2d5b04831fb802e97c30b7fc7cd229bfe8afaa2a1" exitCode=0 Dec 15 12:53:40 crc kubenswrapper[4719]: I1215 12:53:40.405444 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" event={"ID":"611b65c2-c554-4f37-a644-41bcbd27ad46","Type":"ContainerDied","Data":"4cf6a93d4de44f32f710d1a2d5b04831fb802e97c30b7fc7cd229bfe8afaa2a1"} Dec 15 12:53:41 crc kubenswrapper[4719]: I1215 12:53:41.854295 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:53:41 crc kubenswrapper[4719]: I1215 12:53:41.979733 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-ssh-key\") pod \"611b65c2-c554-4f37-a644-41bcbd27ad46\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " Dec 15 12:53:41 crc kubenswrapper[4719]: I1215 12:53:41.979805 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-inventory\") pod \"611b65c2-c554-4f37-a644-41bcbd27ad46\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " Dec 15 12:53:41 crc kubenswrapper[4719]: I1215 12:53:41.979913 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlnd9\" (UniqueName: \"kubernetes.io/projected/611b65c2-c554-4f37-a644-41bcbd27ad46-kube-api-access-vlnd9\") pod \"611b65c2-c554-4f37-a644-41bcbd27ad46\" (UID: \"611b65c2-c554-4f37-a644-41bcbd27ad46\") " Dec 15 12:53:41 crc kubenswrapper[4719]: I1215 12:53:41.991074 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/611b65c2-c554-4f37-a644-41bcbd27ad46-kube-api-access-vlnd9" (OuterVolumeSpecName: "kube-api-access-vlnd9") pod "611b65c2-c554-4f37-a644-41bcbd27ad46" (UID: "611b65c2-c554-4f37-a644-41bcbd27ad46"). InnerVolumeSpecName "kube-api-access-vlnd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.036231 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "611b65c2-c554-4f37-a644-41bcbd27ad46" (UID: "611b65c2-c554-4f37-a644-41bcbd27ad46"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.083834 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.083879 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlnd9\" (UniqueName: \"kubernetes.io/projected/611b65c2-c554-4f37-a644-41bcbd27ad46-kube-api-access-vlnd9\") on node \"crc\" DevicePath \"\"" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.124097 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-inventory" (OuterVolumeSpecName: "inventory") pod "611b65c2-c554-4f37-a644-41bcbd27ad46" (UID: "611b65c2-c554-4f37-a644-41bcbd27ad46"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.187950 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/611b65c2-c554-4f37-a644-41bcbd27ad46-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.425759 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" event={"ID":"611b65c2-c554-4f37-a644-41bcbd27ad46","Type":"ContainerDied","Data":"954a00f9f4b05fb2bb88bce4acb7b6d667f072e71060270580a9e0d6404e6f44"} Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.425798 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.425803 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="954a00f9f4b05fb2bb88bce4acb7b6d667f072e71060270580a9e0d6404e6f44" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.524465 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xhxtk"] Dec 15 12:53:42 crc kubenswrapper[4719]: E1215 12:53:42.524963 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="611b65c2-c554-4f37-a644-41bcbd27ad46" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.525022 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="611b65c2-c554-4f37-a644-41bcbd27ad46" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.525243 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="611b65c2-c554-4f37-a644-41bcbd27ad46" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.528979 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.532197 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.532360 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.532643 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.534014 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.535671 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xhxtk"] Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.697030 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xhxtk\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.697150 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xhxtk\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.697228 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9bpn\" (UniqueName: \"kubernetes.io/projected/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-kube-api-access-n9bpn\") pod \"ssh-known-hosts-edpm-deployment-xhxtk\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.799145 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xhxtk\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.799306 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9bpn\" (UniqueName: \"kubernetes.io/projected/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-kube-api-access-n9bpn\") pod \"ssh-known-hosts-edpm-deployment-xhxtk\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.799488 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xhxtk\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.803414 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xhxtk\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.805440 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xhxtk\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.825680 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9bpn\" (UniqueName: \"kubernetes.io/projected/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-kube-api-access-n9bpn\") pod \"ssh-known-hosts-edpm-deployment-xhxtk\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:42 crc kubenswrapper[4719]: I1215 12:53:42.843848 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:43 crc kubenswrapper[4719]: I1215 12:53:43.376003 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xhxtk"] Dec 15 12:53:43 crc kubenswrapper[4719]: W1215 12:53:43.387116 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7b477ac_26a4_4ddb_af83_57d931fe0e5a.slice/crio-e7fcb8e6c71cc90ad0f7c805cec814056eddc6ac4d55fb6ed121df536607c415 WatchSource:0}: Error finding container e7fcb8e6c71cc90ad0f7c805cec814056eddc6ac4d55fb6ed121df536607c415: Status 404 returned error can't find the container with id e7fcb8e6c71cc90ad0f7c805cec814056eddc6ac4d55fb6ed121df536607c415 Dec 15 12:53:43 crc kubenswrapper[4719]: I1215 12:53:43.391302 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 12:53:43 crc kubenswrapper[4719]: I1215 12:53:43.434260 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" event={"ID":"c7b477ac-26a4-4ddb-af83-57d931fe0e5a","Type":"ContainerStarted","Data":"e7fcb8e6c71cc90ad0f7c805cec814056eddc6ac4d55fb6ed121df536607c415"} Dec 15 12:53:44 crc kubenswrapper[4719]: I1215 12:53:44.445681 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" event={"ID":"c7b477ac-26a4-4ddb-af83-57d931fe0e5a","Type":"ContainerStarted","Data":"b9c3d8ecce621d3db5e227b314d99bdcfb041b5542bcb197b038423558a5c1b3"} Dec 15 12:53:44 crc kubenswrapper[4719]: I1215 12:53:44.466881 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" podStartSLOduration=1.8525518170000002 podStartE2EDuration="2.466841272s" podCreationTimestamp="2025-12-15 12:53:42 +0000 UTC" firstStartedPulling="2025-12-15 12:53:43.391060694 +0000 UTC m=+2184.333353724" lastFinishedPulling="2025-12-15 12:53:44.005350149 +0000 UTC m=+2184.947643179" observedRunningTime="2025-12-15 12:53:44.458567352 +0000 UTC m=+2185.400860392" watchObservedRunningTime="2025-12-15 12:53:44.466841272 +0000 UTC m=+2185.409134312" Dec 15 12:53:51 crc kubenswrapper[4719]: I1215 12:53:51.519728 4719 generic.go:334] "Generic (PLEG): container finished" podID="c7b477ac-26a4-4ddb-af83-57d931fe0e5a" containerID="b9c3d8ecce621d3db5e227b314d99bdcfb041b5542bcb197b038423558a5c1b3" exitCode=0 Dec 15 12:53:51 crc kubenswrapper[4719]: I1215 12:53:51.519807 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" event={"ID":"c7b477ac-26a4-4ddb-af83-57d931fe0e5a","Type":"ContainerDied","Data":"b9c3d8ecce621d3db5e227b314d99bdcfb041b5542bcb197b038423558a5c1b3"} Dec 15 12:53:52 crc kubenswrapper[4719]: I1215 12:53:52.976985 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.097618 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-inventory-0\") pod \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.097758 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9bpn\" (UniqueName: \"kubernetes.io/projected/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-kube-api-access-n9bpn\") pod \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.097799 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-ssh-key-openstack-edpm-ipam\") pod \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\" (UID: \"c7b477ac-26a4-4ddb-af83-57d931fe0e5a\") " Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.103684 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-kube-api-access-n9bpn" (OuterVolumeSpecName: "kube-api-access-n9bpn") pod "c7b477ac-26a4-4ddb-af83-57d931fe0e5a" (UID: "c7b477ac-26a4-4ddb-af83-57d931fe0e5a"). InnerVolumeSpecName "kube-api-access-n9bpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.127400 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c7b477ac-26a4-4ddb-af83-57d931fe0e5a" (UID: "c7b477ac-26a4-4ddb-af83-57d931fe0e5a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.144113 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "c7b477ac-26a4-4ddb-af83-57d931fe0e5a" (UID: "c7b477ac-26a4-4ddb-af83-57d931fe0e5a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.200648 4719 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.200837 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9bpn\" (UniqueName: \"kubernetes.io/projected/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-kube-api-access-n9bpn\") on node \"crc\" DevicePath \"\"" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.200991 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7b477ac-26a4-4ddb-af83-57d931fe0e5a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.546296 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" event={"ID":"c7b477ac-26a4-4ddb-af83-57d931fe0e5a","Type":"ContainerDied","Data":"e7fcb8e6c71cc90ad0f7c805cec814056eddc6ac4d55fb6ed121df536607c415"} Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.546339 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7fcb8e6c71cc90ad0f7c805cec814056eddc6ac4d55fb6ed121df536607c415" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.546461 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xhxtk" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.629786 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr"] Dec 15 12:53:53 crc kubenswrapper[4719]: E1215 12:53:53.630663 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7b477ac-26a4-4ddb-af83-57d931fe0e5a" containerName="ssh-known-hosts-edpm-deployment" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.630915 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7b477ac-26a4-4ddb-af83-57d931fe0e5a" containerName="ssh-known-hosts-edpm-deployment" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.631220 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7b477ac-26a4-4ddb-af83-57d931fe0e5a" containerName="ssh-known-hosts-edpm-deployment" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.631947 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.634619 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.634735 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.634958 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.635008 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.640502 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr"] Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.721339 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kfrcr\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.723371 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kfrcr\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.724280 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf4sn\" (UniqueName: \"kubernetes.io/projected/57d95048-e4af-466c-84ff-f2d00d3b5f58-kube-api-access-zf4sn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kfrcr\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.826946 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kfrcr\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.827132 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kfrcr\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.827173 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf4sn\" (UniqueName: \"kubernetes.io/projected/57d95048-e4af-466c-84ff-f2d00d3b5f58-kube-api-access-zf4sn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kfrcr\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.831763 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kfrcr\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.831952 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kfrcr\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.850047 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf4sn\" (UniqueName: \"kubernetes.io/projected/57d95048-e4af-466c-84ff-f2d00d3b5f58-kube-api-access-zf4sn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-kfrcr\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:53 crc kubenswrapper[4719]: I1215 12:53:53.959709 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:53:54 crc kubenswrapper[4719]: I1215 12:53:54.509806 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr"] Dec 15 12:53:54 crc kubenswrapper[4719]: I1215 12:53:54.556240 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" event={"ID":"57d95048-e4af-466c-84ff-f2d00d3b5f58","Type":"ContainerStarted","Data":"59df4640d4478bd2bbbf6396ce2e3dc5e492568d545bc4fac0b68a50d9271ea6"} Dec 15 12:53:55 crc kubenswrapper[4719]: I1215 12:53:55.565179 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" event={"ID":"57d95048-e4af-466c-84ff-f2d00d3b5f58","Type":"ContainerStarted","Data":"fe5dac249e31908e99a106b6b74b120bb9dcfbd7b5c022d7653c144da8c6093f"} Dec 15 12:53:55 crc kubenswrapper[4719]: I1215 12:53:55.590603 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" podStartSLOduration=2.143107143 podStartE2EDuration="2.590584667s" podCreationTimestamp="2025-12-15 12:53:53 +0000 UTC" firstStartedPulling="2025-12-15 12:53:54.513144766 +0000 UTC m=+2195.455437806" lastFinishedPulling="2025-12-15 12:53:54.9606223 +0000 UTC m=+2195.902915330" observedRunningTime="2025-12-15 12:53:55.58466019 +0000 UTC m=+2196.526953220" watchObservedRunningTime="2025-12-15 12:53:55.590584667 +0000 UTC m=+2196.532877697" Dec 15 12:54:04 crc kubenswrapper[4719]: I1215 12:54:04.667720 4719 generic.go:334] "Generic (PLEG): container finished" podID="57d95048-e4af-466c-84ff-f2d00d3b5f58" containerID="fe5dac249e31908e99a106b6b74b120bb9dcfbd7b5c022d7653c144da8c6093f" exitCode=0 Dec 15 12:54:04 crc kubenswrapper[4719]: I1215 12:54:04.667812 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" event={"ID":"57d95048-e4af-466c-84ff-f2d00d3b5f58","Type":"ContainerDied","Data":"fe5dac249e31908e99a106b6b74b120bb9dcfbd7b5c022d7653c144da8c6093f"} Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.056075 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.166113 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-inventory\") pod \"57d95048-e4af-466c-84ff-f2d00d3b5f58\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.166203 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-ssh-key\") pod \"57d95048-e4af-466c-84ff-f2d00d3b5f58\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.166276 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf4sn\" (UniqueName: \"kubernetes.io/projected/57d95048-e4af-466c-84ff-f2d00d3b5f58-kube-api-access-zf4sn\") pod \"57d95048-e4af-466c-84ff-f2d00d3b5f58\" (UID: \"57d95048-e4af-466c-84ff-f2d00d3b5f58\") " Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.171497 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57d95048-e4af-466c-84ff-f2d00d3b5f58-kube-api-access-zf4sn" (OuterVolumeSpecName: "kube-api-access-zf4sn") pod "57d95048-e4af-466c-84ff-f2d00d3b5f58" (UID: "57d95048-e4af-466c-84ff-f2d00d3b5f58"). InnerVolumeSpecName "kube-api-access-zf4sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.203294 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-inventory" (OuterVolumeSpecName: "inventory") pod "57d95048-e4af-466c-84ff-f2d00d3b5f58" (UID: "57d95048-e4af-466c-84ff-f2d00d3b5f58"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.217336 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "57d95048-e4af-466c-84ff-f2d00d3b5f58" (UID: "57d95048-e4af-466c-84ff-f2d00d3b5f58"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.268540 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.268589 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf4sn\" (UniqueName: \"kubernetes.io/projected/57d95048-e4af-466c-84ff-f2d00d3b5f58-kube-api-access-zf4sn\") on node \"crc\" DevicePath \"\"" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.268610 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57d95048-e4af-466c-84ff-f2d00d3b5f58-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.688367 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" event={"ID":"57d95048-e4af-466c-84ff-f2d00d3b5f58","Type":"ContainerDied","Data":"59df4640d4478bd2bbbf6396ce2e3dc5e492568d545bc4fac0b68a50d9271ea6"} Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.688459 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59df4640d4478bd2bbbf6396ce2e3dc5e492568d545bc4fac0b68a50d9271ea6" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.688492 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-kfrcr" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.783008 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk"] Dec 15 12:54:06 crc kubenswrapper[4719]: E1215 12:54:06.783551 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d95048-e4af-466c-84ff-f2d00d3b5f58" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.783573 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d95048-e4af-466c-84ff-f2d00d3b5f58" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.783792 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="57d95048-e4af-466c-84ff-f2d00d3b5f58" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.784607 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.787299 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.787528 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.787641 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.787760 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.793150 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk"] Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.880802 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.880914 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.881086 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62xp\" (UniqueName: \"kubernetes.io/projected/f1474b5e-e320-4f22-97ae-729dbdd97939-kube-api-access-m62xp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.982799 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62xp\" (UniqueName: \"kubernetes.io/projected/f1474b5e-e320-4f22-97ae-729dbdd97939-kube-api-access-m62xp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.982955 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.983012 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.986996 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.989392 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:06 crc kubenswrapper[4719]: I1215 12:54:06.999269 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62xp\" (UniqueName: \"kubernetes.io/projected/f1474b5e-e320-4f22-97ae-729dbdd97939-kube-api-access-m62xp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:07 crc kubenswrapper[4719]: I1215 12:54:07.102219 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:07 crc kubenswrapper[4719]: I1215 12:54:07.621752 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk"] Dec 15 12:54:07 crc kubenswrapper[4719]: I1215 12:54:07.696642 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" event={"ID":"f1474b5e-e320-4f22-97ae-729dbdd97939","Type":"ContainerStarted","Data":"b8fead1f3e55d77a7e5980fe20060bbf4619a6d0d50d40d11dfee7bc5fc95d35"} Dec 15 12:54:08 crc kubenswrapper[4719]: I1215 12:54:08.706169 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" event={"ID":"f1474b5e-e320-4f22-97ae-729dbdd97939","Type":"ContainerStarted","Data":"51178deab2f2597478c802c3f1e208b5c8ec4d7f37a5256789d0973c181532f1"} Dec 15 12:54:08 crc kubenswrapper[4719]: I1215 12:54:08.740469 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" podStartSLOduration=2.289890277 podStartE2EDuration="2.740447237s" podCreationTimestamp="2025-12-15 12:54:06 +0000 UTC" firstStartedPulling="2025-12-15 12:54:07.626234012 +0000 UTC m=+2208.568527042" lastFinishedPulling="2025-12-15 12:54:08.076790942 +0000 UTC m=+2209.019084002" observedRunningTime="2025-12-15 12:54:08.732322426 +0000 UTC m=+2209.674615476" watchObservedRunningTime="2025-12-15 12:54:08.740447237 +0000 UTC m=+2209.682740277" Dec 15 12:54:17 crc kubenswrapper[4719]: I1215 12:54:17.788629 4719 generic.go:334] "Generic (PLEG): container finished" podID="f1474b5e-e320-4f22-97ae-729dbdd97939" containerID="51178deab2f2597478c802c3f1e208b5c8ec4d7f37a5256789d0973c181532f1" exitCode=0 Dec 15 12:54:17 crc kubenswrapper[4719]: I1215 12:54:17.788872 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" event={"ID":"f1474b5e-e320-4f22-97ae-729dbdd97939","Type":"ContainerDied","Data":"51178deab2f2597478c802c3f1e208b5c8ec4d7f37a5256789d0973c181532f1"} Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.197255 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.325916 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m62xp\" (UniqueName: \"kubernetes.io/projected/f1474b5e-e320-4f22-97ae-729dbdd97939-kube-api-access-m62xp\") pod \"f1474b5e-e320-4f22-97ae-729dbdd97939\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.326509 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-ssh-key\") pod \"f1474b5e-e320-4f22-97ae-729dbdd97939\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.326579 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-inventory\") pod \"f1474b5e-e320-4f22-97ae-729dbdd97939\" (UID: \"f1474b5e-e320-4f22-97ae-729dbdd97939\") " Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.335223 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1474b5e-e320-4f22-97ae-729dbdd97939-kube-api-access-m62xp" (OuterVolumeSpecName: "kube-api-access-m62xp") pod "f1474b5e-e320-4f22-97ae-729dbdd97939" (UID: "f1474b5e-e320-4f22-97ae-729dbdd97939"). InnerVolumeSpecName "kube-api-access-m62xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.354367 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-inventory" (OuterVolumeSpecName: "inventory") pod "f1474b5e-e320-4f22-97ae-729dbdd97939" (UID: "f1474b5e-e320-4f22-97ae-729dbdd97939"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.363791 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f1474b5e-e320-4f22-97ae-729dbdd97939" (UID: "f1474b5e-e320-4f22-97ae-729dbdd97939"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.429139 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m62xp\" (UniqueName: \"kubernetes.io/projected/f1474b5e-e320-4f22-97ae-729dbdd97939-kube-api-access-m62xp\") on node \"crc\" DevicePath \"\"" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.429174 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.429184 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1474b5e-e320-4f22-97ae-729dbdd97939-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.814768 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" event={"ID":"f1474b5e-e320-4f22-97ae-729dbdd97939","Type":"ContainerDied","Data":"b8fead1f3e55d77a7e5980fe20060bbf4619a6d0d50d40d11dfee7bc5fc95d35"} Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.814835 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8fead1f3e55d77a7e5980fe20060bbf4619a6d0d50d40d11dfee7bc5fc95d35" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.814970 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.917568 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s"] Dec 15 12:54:19 crc kubenswrapper[4719]: E1215 12:54:19.918023 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1474b5e-e320-4f22-97ae-729dbdd97939" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.918044 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1474b5e-e320-4f22-97ae-729dbdd97939" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.918262 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1474b5e-e320-4f22-97ae-729dbdd97939" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.919013 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.924333 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.925484 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.928776 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.929817 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.929880 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.931012 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.931333 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.934766 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 15 12:54:19 crc kubenswrapper[4719]: I1215 12:54:19.938378 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s"] Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.046368 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.046648 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.046766 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.046888 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.047013 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.047113 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.047191 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.047310 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.047388 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.047478 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.047559 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.047632 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c66m7\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-kube-api-access-c66m7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.047728 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.047824 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149342 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149384 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149412 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149441 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149514 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149555 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149595 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149613 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c66m7\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-kube-api-access-c66m7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149629 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149656 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149674 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149696 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149780 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.149805 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.153494 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.153762 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.154203 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.154603 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.154685 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.155530 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.156844 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.157414 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.159004 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.159306 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.159658 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.160113 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.161080 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.170343 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c66m7\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-kube-api-access-c66m7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.238409 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:54:20 crc kubenswrapper[4719]: I1215 12:54:20.825151 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s"] Dec 15 12:54:21 crc kubenswrapper[4719]: I1215 12:54:21.357121 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:54:21 crc kubenswrapper[4719]: I1215 12:54:21.357194 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:54:21 crc kubenswrapper[4719]: I1215 12:54:21.840415 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" event={"ID":"71b47db7-3079-4990-8a37-b02ea6f0368b","Type":"ContainerStarted","Data":"ced1d0fb6485a80e8ffc54c62bc6fba6bd3eb2e586ce4c06a7915369f4f24154"} Dec 15 12:54:21 crc kubenswrapper[4719]: I1215 12:54:21.840772 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" event={"ID":"71b47db7-3079-4990-8a37-b02ea6f0368b","Type":"ContainerStarted","Data":"6c39875516cb8a4d0c37200843058912905d90f6f26e9ae54efada5bfb67af61"} Dec 15 12:54:21 crc kubenswrapper[4719]: I1215 12:54:21.880576 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" podStartSLOduration=2.393750179 podStartE2EDuration="2.880544349s" podCreationTimestamp="2025-12-15 12:54:19 +0000 UTC" firstStartedPulling="2025-12-15 12:54:20.828285253 +0000 UTC m=+2221.770578293" lastFinishedPulling="2025-12-15 12:54:21.315079423 +0000 UTC m=+2222.257372463" observedRunningTime="2025-12-15 12:54:21.865958327 +0000 UTC m=+2222.808251357" watchObservedRunningTime="2025-12-15 12:54:21.880544349 +0000 UTC m=+2222.822837449" Dec 15 12:54:22 crc kubenswrapper[4719]: E1215 12:54:22.680103 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice/crio-b8fead1f3e55d77a7e5980fe20060bbf4619a6d0d50d40d11dfee7bc5fc95d35\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice\": RecentStats: unable to find data in memory cache]" Dec 15 12:54:32 crc kubenswrapper[4719]: E1215 12:54:32.941791 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice/crio-b8fead1f3e55d77a7e5980fe20060bbf4619a6d0d50d40d11dfee7bc5fc95d35\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice\": RecentStats: unable to find data in memory cache]" Dec 15 12:54:43 crc kubenswrapper[4719]: E1215 12:54:43.184938 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice/crio-b8fead1f3e55d77a7e5980fe20060bbf4619a6d0d50d40d11dfee7bc5fc95d35\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice\": RecentStats: unable to find data in memory cache]" Dec 15 12:54:51 crc kubenswrapper[4719]: I1215 12:54:51.357403 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:54:51 crc kubenswrapper[4719]: I1215 12:54:51.357982 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:54:53 crc kubenswrapper[4719]: E1215 12:54:53.403362 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice/crio-b8fead1f3e55d77a7e5980fe20060bbf4619a6d0d50d40d11dfee7bc5fc95d35\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice\": RecentStats: unable to find data in memory cache]" Dec 15 12:54:59 crc kubenswrapper[4719]: I1215 12:54:59.165996 4719 generic.go:334] "Generic (PLEG): container finished" podID="71b47db7-3079-4990-8a37-b02ea6f0368b" containerID="ced1d0fb6485a80e8ffc54c62bc6fba6bd3eb2e586ce4c06a7915369f4f24154" exitCode=0 Dec 15 12:54:59 crc kubenswrapper[4719]: I1215 12:54:59.166048 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" event={"ID":"71b47db7-3079-4990-8a37-b02ea6f0368b","Type":"ContainerDied","Data":"ced1d0fb6485a80e8ffc54c62bc6fba6bd3eb2e586ce4c06a7915369f4f24154"} Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.648809 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721015 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721068 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-inventory\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721105 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721173 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ovn-combined-ca-bundle\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721190 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-nova-combined-ca-bundle\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721210 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-bootstrap-combined-ca-bundle\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721251 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-libvirt-combined-ca-bundle\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721292 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-repo-setup-combined-ca-bundle\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721329 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c66m7\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-kube-api-access-c66m7\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721368 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721456 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-neutron-metadata-combined-ca-bundle\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721492 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-telemetry-combined-ca-bundle\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721518 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ssh-key\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.721534 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"71b47db7-3079-4990-8a37-b02ea6f0368b\" (UID: \"71b47db7-3079-4990-8a37-b02ea6f0368b\") " Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.735732 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.742002 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.744397 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.751054 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-kube-api-access-c66m7" (OuterVolumeSpecName: "kube-api-access-c66m7") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "kube-api-access-c66m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.751155 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.759167 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.759214 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.759322 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.759577 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.762993 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.776085 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.776206 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.798473 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.807180 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-inventory" (OuterVolumeSpecName: "inventory") pod "71b47db7-3079-4990-8a37-b02ea6f0368b" (UID: "71b47db7-3079-4990-8a37-b02ea6f0368b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823548 4719 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823579 4719 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823590 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823600 4719 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823613 4719 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823622 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823632 4719 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823642 4719 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823651 4719 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823659 4719 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823668 4719 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823676 4719 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71b47db7-3079-4990-8a37-b02ea6f0368b-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823686 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c66m7\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-kube-api-access-c66m7\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:00 crc kubenswrapper[4719]: I1215 12:55:00.823695 4719 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/71b47db7-3079-4990-8a37-b02ea6f0368b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.187720 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" event={"ID":"71b47db7-3079-4990-8a37-b02ea6f0368b","Type":"ContainerDied","Data":"6c39875516cb8a4d0c37200843058912905d90f6f26e9ae54efada5bfb67af61"} Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.188175 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c39875516cb8a4d0c37200843058912905d90f6f26e9ae54efada5bfb67af61" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.187971 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.321606 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7"] Dec 15 12:55:01 crc kubenswrapper[4719]: E1215 12:55:01.322211 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b47db7-3079-4990-8a37-b02ea6f0368b" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.322243 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b47db7-3079-4990-8a37-b02ea6f0368b" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.322525 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b47db7-3079-4990-8a37-b02ea6f0368b" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.323523 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.326316 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.326470 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.326572 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.327532 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.327832 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.335948 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7"] Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.437198 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.437285 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.437322 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgw2r\" (UniqueName: \"kubernetes.io/projected/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-kube-api-access-dgw2r\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.437433 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.437491 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.538673 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.538739 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.538766 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgw2r\" (UniqueName: \"kubernetes.io/projected/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-kube-api-access-dgw2r\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.539520 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.540641 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.540770 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.543353 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.543557 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.543723 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.566110 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgw2r\" (UniqueName: \"kubernetes.io/projected/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-kube-api-access-dgw2r\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nrfr7\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:01 crc kubenswrapper[4719]: I1215 12:55:01.639535 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:55:02 crc kubenswrapper[4719]: I1215 12:55:02.160184 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7"] Dec 15 12:55:02 crc kubenswrapper[4719]: I1215 12:55:02.200809 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" event={"ID":"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3","Type":"ContainerStarted","Data":"b235198257f2a34bd8d8793a8a5608c978af76aa33acc5c9829f0db563079e79"} Dec 15 12:55:03 crc kubenswrapper[4719]: I1215 12:55:03.214288 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" event={"ID":"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3","Type":"ContainerStarted","Data":"d1925f930b632d6ef082857849d686404961f739e12a30f41c014dff54f7aabc"} Dec 15 12:55:03 crc kubenswrapper[4719]: I1215 12:55:03.264871 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" podStartSLOduration=1.797023341 podStartE2EDuration="2.264832984s" podCreationTimestamp="2025-12-15 12:55:01 +0000 UTC" firstStartedPulling="2025-12-15 12:55:02.171833317 +0000 UTC m=+2263.114126347" lastFinishedPulling="2025-12-15 12:55:02.63964291 +0000 UTC m=+2263.581935990" observedRunningTime="2025-12-15 12:55:03.246614411 +0000 UTC m=+2264.188907451" watchObservedRunningTime="2025-12-15 12:55:03.264832984 +0000 UTC m=+2264.207126024" Dec 15 12:55:03 crc kubenswrapper[4719]: E1215 12:55:03.680261 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice/crio-b8fead1f3e55d77a7e5980fe20060bbf4619a6d0d50d40d11dfee7bc5fc95d35\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice\": RecentStats: unable to find data in memory cache]" Dec 15 12:55:13 crc kubenswrapper[4719]: E1215 12:55:13.960058 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1474b5e_e320_4f22_97ae_729dbdd97939.slice/crio-b8fead1f3e55d77a7e5980fe20060bbf4619a6d0d50d40d11dfee7bc5fc95d35\": RecentStats: unable to find data in memory cache]" Dec 15 12:55:21 crc kubenswrapper[4719]: I1215 12:55:21.356600 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 12:55:21 crc kubenswrapper[4719]: I1215 12:55:21.357135 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 12:55:21 crc kubenswrapper[4719]: I1215 12:55:21.357176 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 12:55:21 crc kubenswrapper[4719]: I1215 12:55:21.357893 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 12:55:21 crc kubenswrapper[4719]: I1215 12:55:21.357939 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" gracePeriod=600 Dec 15 12:55:21 crc kubenswrapper[4719]: E1215 12:55:21.476033 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:55:22 crc kubenswrapper[4719]: I1215 12:55:22.384244 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" exitCode=0 Dec 15 12:55:22 crc kubenswrapper[4719]: I1215 12:55:22.384291 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58"} Dec 15 12:55:22 crc kubenswrapper[4719]: I1215 12:55:22.384381 4719 scope.go:117] "RemoveContainer" containerID="23da24f9e77dd91d08a2fcb3b8667d259f3c0a3ecdb0953729476936101d7499" Dec 15 12:55:22 crc kubenswrapper[4719]: I1215 12:55:22.385073 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:55:22 crc kubenswrapper[4719]: E1215 12:55:22.385463 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:55:34 crc kubenswrapper[4719]: I1215 12:55:34.476261 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:55:34 crc kubenswrapper[4719]: E1215 12:55:34.477024 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:55:48 crc kubenswrapper[4719]: I1215 12:55:48.476724 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:55:48 crc kubenswrapper[4719]: E1215 12:55:48.477520 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:55:59 crc kubenswrapper[4719]: I1215 12:55:59.481443 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:55:59 crc kubenswrapper[4719]: E1215 12:55:59.482236 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:56:13 crc kubenswrapper[4719]: I1215 12:56:13.477082 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:56:13 crc kubenswrapper[4719]: E1215 12:56:13.477912 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:56:15 crc kubenswrapper[4719]: I1215 12:56:15.895112 4719 generic.go:334] "Generic (PLEG): container finished" podID="87e0fa54-5d1a-4e53-aef8-257af1ebd4d3" containerID="d1925f930b632d6ef082857849d686404961f739e12a30f41c014dff54f7aabc" exitCode=0 Dec 15 12:56:15 crc kubenswrapper[4719]: I1215 12:56:15.895182 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" event={"ID":"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3","Type":"ContainerDied","Data":"d1925f930b632d6ef082857849d686404961f739e12a30f41c014dff54f7aabc"} Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.288754 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.452543 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovncontroller-config-0\") pod \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.452669 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgw2r\" (UniqueName: \"kubernetes.io/projected/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-kube-api-access-dgw2r\") pod \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.452755 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovn-combined-ca-bundle\") pod \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.452818 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-inventory\") pod \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.452920 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ssh-key\") pod \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\" (UID: \"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3\") " Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.470709 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "87e0fa54-5d1a-4e53-aef8-257af1ebd4d3" (UID: "87e0fa54-5d1a-4e53-aef8-257af1ebd4d3"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.470754 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-kube-api-access-dgw2r" (OuterVolumeSpecName: "kube-api-access-dgw2r") pod "87e0fa54-5d1a-4e53-aef8-257af1ebd4d3" (UID: "87e0fa54-5d1a-4e53-aef8-257af1ebd4d3"). InnerVolumeSpecName "kube-api-access-dgw2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.479231 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "87e0fa54-5d1a-4e53-aef8-257af1ebd4d3" (UID: "87e0fa54-5d1a-4e53-aef8-257af1ebd4d3"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.484155 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "87e0fa54-5d1a-4e53-aef8-257af1ebd4d3" (UID: "87e0fa54-5d1a-4e53-aef8-257af1ebd4d3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.499942 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-inventory" (OuterVolumeSpecName: "inventory") pod "87e0fa54-5d1a-4e53-aef8-257af1ebd4d3" (UID: "87e0fa54-5d1a-4e53-aef8-257af1ebd4d3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.555064 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.555119 4719 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.555138 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgw2r\" (UniqueName: \"kubernetes.io/projected/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-kube-api-access-dgw2r\") on node \"crc\" DevicePath \"\"" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.555150 4719 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.555164 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87e0fa54-5d1a-4e53-aef8-257af1ebd4d3-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.911682 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" event={"ID":"87e0fa54-5d1a-4e53-aef8-257af1ebd4d3","Type":"ContainerDied","Data":"b235198257f2a34bd8d8793a8a5608c978af76aa33acc5c9829f0db563079e79"} Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.912050 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b235198257f2a34bd8d8793a8a5608c978af76aa33acc5c9829f0db563079e79" Dec 15 12:56:17 crc kubenswrapper[4719]: I1215 12:56:17.911700 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nrfr7" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.073277 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg"] Dec 15 12:56:18 crc kubenswrapper[4719]: E1215 12:56:18.073715 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e0fa54-5d1a-4e53-aef8-257af1ebd4d3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.073743 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e0fa54-5d1a-4e53-aef8-257af1ebd4d3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.074008 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e0fa54-5d1a-4e53-aef8-257af1ebd4d3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.074719 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.080462 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.080534 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.080672 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.080755 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.080845 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.081526 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.087083 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg"] Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.168435 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6zqz\" (UniqueName: \"kubernetes.io/projected/39093fe3-9551-4d58-9cd0-607f1694a71c-kube-api-access-w6zqz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.168557 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.168589 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.168713 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.169019 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.169050 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.271257 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.272351 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.272448 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6zqz\" (UniqueName: \"kubernetes.io/projected/39093fe3-9551-4d58-9cd0-607f1694a71c-kube-api-access-w6zqz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.272505 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.272552 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.272603 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.276169 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.276245 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.276579 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.276622 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.295762 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.309742 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6zqz\" (UniqueName: \"kubernetes.io/projected/39093fe3-9551-4d58-9cd0-607f1694a71c-kube-api-access-w6zqz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.400522 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.752413 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg"] Dec 15 12:56:18 crc kubenswrapper[4719]: I1215 12:56:18.920783 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" event={"ID":"39093fe3-9551-4d58-9cd0-607f1694a71c","Type":"ContainerStarted","Data":"58c033d1bd9d1cd5476f550c8a75cd83869a2f426e4f98fd975978b705f9fbc3"} Dec 15 12:56:19 crc kubenswrapper[4719]: I1215 12:56:19.929228 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" event={"ID":"39093fe3-9551-4d58-9cd0-607f1694a71c","Type":"ContainerStarted","Data":"60196972f7b513ec73d6a0238c7bd48e55b759f5b677743b44672276168be8f0"} Dec 15 12:56:19 crc kubenswrapper[4719]: I1215 12:56:19.954932 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" podStartSLOduration=1.407860927 podStartE2EDuration="1.954901232s" podCreationTimestamp="2025-12-15 12:56:18 +0000 UTC" firstStartedPulling="2025-12-15 12:56:18.757672427 +0000 UTC m=+2339.699965447" lastFinishedPulling="2025-12-15 12:56:19.304712722 +0000 UTC m=+2340.247005752" observedRunningTime="2025-12-15 12:56:19.945455529 +0000 UTC m=+2340.887748579" watchObservedRunningTime="2025-12-15 12:56:19.954901232 +0000 UTC m=+2340.897194262" Dec 15 12:56:25 crc kubenswrapper[4719]: I1215 12:56:25.478080 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:56:25 crc kubenswrapper[4719]: E1215 12:56:25.479000 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:56:37 crc kubenswrapper[4719]: I1215 12:56:37.475939 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:56:37 crc kubenswrapper[4719]: E1215 12:56:37.476739 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:56:50 crc kubenswrapper[4719]: I1215 12:56:50.476286 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:56:50 crc kubenswrapper[4719]: E1215 12:56:50.477082 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:57:02 crc kubenswrapper[4719]: I1215 12:57:02.475900 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:57:02 crc kubenswrapper[4719]: E1215 12:57:02.476687 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:57:16 crc kubenswrapper[4719]: E1215 12:57:16.885817 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39093fe3_9551_4d58_9cd0_607f1694a71c.slice/crio-60196972f7b513ec73d6a0238c7bd48e55b759f5b677743b44672276168be8f0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39093fe3_9551_4d58_9cd0_607f1694a71c.slice/crio-conmon-60196972f7b513ec73d6a0238c7bd48e55b759f5b677743b44672276168be8f0.scope\": RecentStats: unable to find data in memory cache]" Dec 15 12:57:17 crc kubenswrapper[4719]: I1215 12:57:17.438126 4719 generic.go:334] "Generic (PLEG): container finished" podID="39093fe3-9551-4d58-9cd0-607f1694a71c" containerID="60196972f7b513ec73d6a0238c7bd48e55b759f5b677743b44672276168be8f0" exitCode=0 Dec 15 12:57:17 crc kubenswrapper[4719]: I1215 12:57:17.438171 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" event={"ID":"39093fe3-9551-4d58-9cd0-607f1694a71c","Type":"ContainerDied","Data":"60196972f7b513ec73d6a0238c7bd48e55b759f5b677743b44672276168be8f0"} Dec 15 12:57:17 crc kubenswrapper[4719]: I1215 12:57:17.479391 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:57:17 crc kubenswrapper[4719]: E1215 12:57:17.480466 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:57:18 crc kubenswrapper[4719]: I1215 12:57:18.885758 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:57:18 crc kubenswrapper[4719]: I1215 12:57:18.982716 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-metadata-combined-ca-bundle\") pod \"39093fe3-9551-4d58-9cd0-607f1694a71c\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " Dec 15 12:57:18 crc kubenswrapper[4719]: I1215 12:57:18.982763 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6zqz\" (UniqueName: \"kubernetes.io/projected/39093fe3-9551-4d58-9cd0-607f1694a71c-kube-api-access-w6zqz\") pod \"39093fe3-9551-4d58-9cd0-607f1694a71c\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " Dec 15 12:57:18 crc kubenswrapper[4719]: I1215 12:57:18.982820 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-ssh-key\") pod \"39093fe3-9551-4d58-9cd0-607f1694a71c\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " Dec 15 12:57:18 crc kubenswrapper[4719]: I1215 12:57:18.982937 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-inventory\") pod \"39093fe3-9551-4d58-9cd0-607f1694a71c\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " Dec 15 12:57:18 crc kubenswrapper[4719]: I1215 12:57:18.982960 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-nova-metadata-neutron-config-0\") pod \"39093fe3-9551-4d58-9cd0-607f1694a71c\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " Dec 15 12:57:18 crc kubenswrapper[4719]: I1215 12:57:18.982980 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"39093fe3-9551-4d58-9cd0-607f1694a71c\" (UID: \"39093fe3-9551-4d58-9cd0-607f1694a71c\") " Dec 15 12:57:18 crc kubenswrapper[4719]: I1215 12:57:18.987755 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39093fe3-9551-4d58-9cd0-607f1694a71c-kube-api-access-w6zqz" (OuterVolumeSpecName: "kube-api-access-w6zqz") pod "39093fe3-9551-4d58-9cd0-607f1694a71c" (UID: "39093fe3-9551-4d58-9cd0-607f1694a71c"). InnerVolumeSpecName "kube-api-access-w6zqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 12:57:18 crc kubenswrapper[4719]: I1215 12:57:18.988207 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "39093fe3-9551-4d58-9cd0-607f1694a71c" (UID: "39093fe3-9551-4d58-9cd0-607f1694a71c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.007710 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "39093fe3-9551-4d58-9cd0-607f1694a71c" (UID: "39093fe3-9551-4d58-9cd0-607f1694a71c"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.010952 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "39093fe3-9551-4d58-9cd0-607f1694a71c" (UID: "39093fe3-9551-4d58-9cd0-607f1694a71c"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.012910 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-inventory" (OuterVolumeSpecName: "inventory") pod "39093fe3-9551-4d58-9cd0-607f1694a71c" (UID: "39093fe3-9551-4d58-9cd0-607f1694a71c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.013150 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "39093fe3-9551-4d58-9cd0-607f1694a71c" (UID: "39093fe3-9551-4d58-9cd0-607f1694a71c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.084399 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.084926 4719 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.085006 4719 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.085068 4719 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.085128 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6zqz\" (UniqueName: \"kubernetes.io/projected/39093fe3-9551-4d58-9cd0-607f1694a71c-kube-api-access-w6zqz\") on node \"crc\" DevicePath \"\"" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.085190 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39093fe3-9551-4d58-9cd0-607f1694a71c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.460027 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" event={"ID":"39093fe3-9551-4d58-9cd0-607f1694a71c","Type":"ContainerDied","Data":"58c033d1bd9d1cd5476f550c8a75cd83869a2f426e4f98fd975978b705f9fbc3"} Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.460368 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58c033d1bd9d1cd5476f550c8a75cd83869a2f426e4f98fd975978b705f9fbc3" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.460038 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.652682 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql"] Dec 15 12:57:19 crc kubenswrapper[4719]: E1215 12:57:19.653190 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39093fe3-9551-4d58-9cd0-607f1694a71c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.653208 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="39093fe3-9551-4d58-9cd0-607f1694a71c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.653418 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="39093fe3-9551-4d58-9cd0-607f1694a71c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.654062 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.656945 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.656979 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.656961 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.657178 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.657676 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.669460 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql"] Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.698325 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.698377 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crv59\" (UniqueName: \"kubernetes.io/projected/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-kube-api-access-crv59\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.698485 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.698537 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.698577 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.799688 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.799742 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crv59\" (UniqueName: \"kubernetes.io/projected/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-kube-api-access-crv59\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.799791 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.799820 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.799849 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.805180 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.806425 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.807769 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.808135 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.822336 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crv59\" (UniqueName: \"kubernetes.io/projected/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-kube-api-access-crv59\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:19 crc kubenswrapper[4719]: I1215 12:57:19.970665 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 12:57:20 crc kubenswrapper[4719]: I1215 12:57:20.340521 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql"] Dec 15 12:57:20 crc kubenswrapper[4719]: I1215 12:57:20.474582 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" event={"ID":"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f","Type":"ContainerStarted","Data":"b0a5893a3e1deaf712855224fd680d4368777a4f8132235738611c1a8aa3e20b"} Dec 15 12:57:21 crc kubenswrapper[4719]: I1215 12:57:21.490712 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" event={"ID":"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f","Type":"ContainerStarted","Data":"b28f0e9b50803bf3e3722620e6dc2fc9fb15836f3a4f4fb9f9a1a4189a7ef625"} Dec 15 12:57:21 crc kubenswrapper[4719]: I1215 12:57:21.514739 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" podStartSLOduration=1.860711556 podStartE2EDuration="2.514713493s" podCreationTimestamp="2025-12-15 12:57:19 +0000 UTC" firstStartedPulling="2025-12-15 12:57:20.353252626 +0000 UTC m=+2401.295545646" lastFinishedPulling="2025-12-15 12:57:21.007254543 +0000 UTC m=+2401.949547583" observedRunningTime="2025-12-15 12:57:21.507983785 +0000 UTC m=+2402.450276845" watchObservedRunningTime="2025-12-15 12:57:21.514713493 +0000 UTC m=+2402.457006563" Dec 15 12:57:30 crc kubenswrapper[4719]: I1215 12:57:30.477213 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:57:30 crc kubenswrapper[4719]: E1215 12:57:30.478250 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:57:34 crc kubenswrapper[4719]: I1215 12:57:34.496194 4719 scope.go:117] "RemoveContainer" containerID="fcb34176c6ecc34af54b83018e02a29bc1c92cbacee137068c35dd4f785d43dd" Dec 15 12:57:34 crc kubenswrapper[4719]: I1215 12:57:34.534543 4719 scope.go:117] "RemoveContainer" containerID="9b0235c84a8b2d9cecbcc960017fe094750ce678c3b4facf733cdc2ae355c969" Dec 15 12:57:45 crc kubenswrapper[4719]: I1215 12:57:45.476417 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:57:45 crc kubenswrapper[4719]: E1215 12:57:45.477239 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:57:56 crc kubenswrapper[4719]: I1215 12:57:56.477525 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:57:56 crc kubenswrapper[4719]: E1215 12:57:56.478888 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:58:08 crc kubenswrapper[4719]: I1215 12:58:08.476257 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:58:08 crc kubenswrapper[4719]: E1215 12:58:08.477397 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:58:23 crc kubenswrapper[4719]: I1215 12:58:23.476235 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:58:23 crc kubenswrapper[4719]: E1215 12:58:23.476936 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:58:34 crc kubenswrapper[4719]: I1215 12:58:34.599548 4719 scope.go:117] "RemoveContainer" containerID="d7ab779eb0c04b66195473c3ac28be5e23577b90a47ff160a48774d885aeb8a9" Dec 15 12:58:37 crc kubenswrapper[4719]: I1215 12:58:37.476919 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:58:37 crc kubenswrapper[4719]: E1215 12:58:37.477328 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:58:49 crc kubenswrapper[4719]: I1215 12:58:49.482193 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:58:49 crc kubenswrapper[4719]: E1215 12:58:49.482991 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:59:00 crc kubenswrapper[4719]: I1215 12:59:00.476992 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:59:00 crc kubenswrapper[4719]: E1215 12:59:00.478179 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:59:14 crc kubenswrapper[4719]: I1215 12:59:14.476131 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:59:14 crc kubenswrapper[4719]: E1215 12:59:14.477272 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:59:28 crc kubenswrapper[4719]: I1215 12:59:28.476936 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:59:28 crc kubenswrapper[4719]: E1215 12:59:28.477688 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:59:42 crc kubenswrapper[4719]: I1215 12:59:42.476298 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:59:42 crc kubenswrapper[4719]: E1215 12:59:42.477156 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:59:54 crc kubenswrapper[4719]: I1215 12:59:54.476595 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 12:59:54 crc kubenswrapper[4719]: E1215 12:59:54.477319 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.556196 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gcgw5"] Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.559566 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.588885 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcgw5"] Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.698275 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-utilities\") pod \"redhat-marketplace-gcgw5\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.698681 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99xqt\" (UniqueName: \"kubernetes.io/projected/18116782-c15c-4c02-8907-12b3a8004cf9-kube-api-access-99xqt\") pod \"redhat-marketplace-gcgw5\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.698883 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-catalog-content\") pod \"redhat-marketplace-gcgw5\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.800731 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-utilities\") pod \"redhat-marketplace-gcgw5\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.801056 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99xqt\" (UniqueName: \"kubernetes.io/projected/18116782-c15c-4c02-8907-12b3a8004cf9-kube-api-access-99xqt\") pod \"redhat-marketplace-gcgw5\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.801176 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-catalog-content\") pod \"redhat-marketplace-gcgw5\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.801255 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-utilities\") pod \"redhat-marketplace-gcgw5\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.801456 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-catalog-content\") pod \"redhat-marketplace-gcgw5\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.824019 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99xqt\" (UniqueName: \"kubernetes.io/projected/18116782-c15c-4c02-8907-12b3a8004cf9-kube-api-access-99xqt\") pod \"redhat-marketplace-gcgw5\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:56 crc kubenswrapper[4719]: I1215 12:59:56.896153 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 12:59:57 crc kubenswrapper[4719]: I1215 12:59:57.225918 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcgw5"] Dec 15 12:59:57 crc kubenswrapper[4719]: I1215 12:59:57.923127 4719 generic.go:334] "Generic (PLEG): container finished" podID="18116782-c15c-4c02-8907-12b3a8004cf9" containerID="9d6cdbffa95611910570c1a6448cf1924c18f18801e1a0d09f7f6ab126e90128" exitCode=0 Dec 15 12:59:57 crc kubenswrapper[4719]: I1215 12:59:57.923231 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcgw5" event={"ID":"18116782-c15c-4c02-8907-12b3a8004cf9","Type":"ContainerDied","Data":"9d6cdbffa95611910570c1a6448cf1924c18f18801e1a0d09f7f6ab126e90128"} Dec 15 12:59:57 crc kubenswrapper[4719]: I1215 12:59:57.923403 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcgw5" event={"ID":"18116782-c15c-4c02-8907-12b3a8004cf9","Type":"ContainerStarted","Data":"27fcfcfb55204dd06321886b0401702e4f0932198a131c9f8250c4563eeb3f97"} Dec 15 12:59:57 crc kubenswrapper[4719]: I1215 12:59:57.925700 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 12:59:59 crc kubenswrapper[4719]: I1215 12:59:59.952169 4719 generic.go:334] "Generic (PLEG): container finished" podID="18116782-c15c-4c02-8907-12b3a8004cf9" containerID="308630120fe5812beeded8d7e7513fbea862871194e8e64e261cb245a3abbea2" exitCode=0 Dec 15 12:59:59 crc kubenswrapper[4719]: I1215 12:59:59.952236 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcgw5" event={"ID":"18116782-c15c-4c02-8907-12b3a8004cf9","Type":"ContainerDied","Data":"308630120fe5812beeded8d7e7513fbea862871194e8e64e261cb245a3abbea2"} Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.163228 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29"] Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.164712 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.167794 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.168410 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.177112 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29"] Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.286175 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45mn5\" (UniqueName: \"kubernetes.io/projected/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-kube-api-access-45mn5\") pod \"collect-profiles-29430060-l5g29\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.286263 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-config-volume\") pod \"collect-profiles-29430060-l5g29\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.286481 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-secret-volume\") pod \"collect-profiles-29430060-l5g29\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.388552 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-config-volume\") pod \"collect-profiles-29430060-l5g29\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.388653 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-secret-volume\") pod \"collect-profiles-29430060-l5g29\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.388764 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45mn5\" (UniqueName: \"kubernetes.io/projected/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-kube-api-access-45mn5\") pod \"collect-profiles-29430060-l5g29\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.389967 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-config-volume\") pod \"collect-profiles-29430060-l5g29\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.395897 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-secret-volume\") pod \"collect-profiles-29430060-l5g29\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.414195 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45mn5\" (UniqueName: \"kubernetes.io/projected/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-kube-api-access-45mn5\") pod \"collect-profiles-29430060-l5g29\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.482411 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.963952 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcgw5" event={"ID":"18116782-c15c-4c02-8907-12b3a8004cf9","Type":"ContainerStarted","Data":"b0cf77812a3a9b6f8f4aafd49709c6eb09270fd14098e8d3752722708a8070bd"} Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.968708 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29"] Dec 15 13:00:00 crc kubenswrapper[4719]: I1215 13:00:00.995343 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gcgw5" podStartSLOduration=2.255853764 podStartE2EDuration="4.995307347s" podCreationTimestamp="2025-12-15 12:59:56 +0000 UTC" firstStartedPulling="2025-12-15 12:59:57.925512296 +0000 UTC m=+2558.867805326" lastFinishedPulling="2025-12-15 13:00:00.664965879 +0000 UTC m=+2561.607258909" observedRunningTime="2025-12-15 13:00:00.987471485 +0000 UTC m=+2561.929764525" watchObservedRunningTime="2025-12-15 13:00:00.995307347 +0000 UTC m=+2561.937600387" Dec 15 13:00:01 crc kubenswrapper[4719]: I1215 13:00:01.972971 4719 generic.go:334] "Generic (PLEG): container finished" podID="29f0f318-9ef6-4f33-aec9-10cbaee5d0e7" containerID="8a6c6ec5b74c2ca53223dcc984ead8444d631af836779f9d69a97ad64d648c11" exitCode=0 Dec 15 13:00:01 crc kubenswrapper[4719]: I1215 13:00:01.973073 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" event={"ID":"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7","Type":"ContainerDied","Data":"8a6c6ec5b74c2ca53223dcc984ead8444d631af836779f9d69a97ad64d648c11"} Dec 15 13:00:01 crc kubenswrapper[4719]: I1215 13:00:01.973302 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" event={"ID":"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7","Type":"ContainerStarted","Data":"773b878678cc553d41c2f64b5a181bfe4b783b4da26d0d40ae4bfcdb2ae5fb05"} Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.325569 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.451281 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-secret-volume\") pod \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.451790 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45mn5\" (UniqueName: \"kubernetes.io/projected/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-kube-api-access-45mn5\") pod \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.451960 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-config-volume\") pod \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\" (UID: \"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7\") " Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.452421 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-config-volume" (OuterVolumeSpecName: "config-volume") pod "29f0f318-9ef6-4f33-aec9-10cbaee5d0e7" (UID: "29f0f318-9ef6-4f33-aec9-10cbaee5d0e7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.457732 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-kube-api-access-45mn5" (OuterVolumeSpecName: "kube-api-access-45mn5") pod "29f0f318-9ef6-4f33-aec9-10cbaee5d0e7" (UID: "29f0f318-9ef6-4f33-aec9-10cbaee5d0e7"). InnerVolumeSpecName "kube-api-access-45mn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.458443 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "29f0f318-9ef6-4f33-aec9-10cbaee5d0e7" (UID: "29f0f318-9ef6-4f33-aec9-10cbaee5d0e7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.554045 4719 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.554084 4719 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.554096 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45mn5\" (UniqueName: \"kubernetes.io/projected/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7-kube-api-access-45mn5\") on node \"crc\" DevicePath \"\"" Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.991024 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" event={"ID":"29f0f318-9ef6-4f33-aec9-10cbaee5d0e7","Type":"ContainerDied","Data":"773b878678cc553d41c2f64b5a181bfe4b783b4da26d0d40ae4bfcdb2ae5fb05"} Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.991063 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="773b878678cc553d41c2f64b5a181bfe4b783b4da26d0d40ae4bfcdb2ae5fb05" Dec 15 13:00:03 crc kubenswrapper[4719]: I1215 13:00:03.991089 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29" Dec 15 13:00:04 crc kubenswrapper[4719]: I1215 13:00:04.407223 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb"] Dec 15 13:00:04 crc kubenswrapper[4719]: I1215 13:00:04.420016 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430015-f8vkb"] Dec 15 13:00:05 crc kubenswrapper[4719]: I1215 13:00:05.490340 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b7161e6-9af7-4443-9af7-2d11e0f36d68" path="/var/lib/kubelet/pods/6b7161e6-9af7-4443-9af7-2d11e0f36d68/volumes" Dec 15 13:00:06 crc kubenswrapper[4719]: I1215 13:00:06.476743 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 13:00:06 crc kubenswrapper[4719]: E1215 13:00:06.477106 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:00:06 crc kubenswrapper[4719]: I1215 13:00:06.897095 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 13:00:06 crc kubenswrapper[4719]: I1215 13:00:06.897145 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 13:00:06 crc kubenswrapper[4719]: I1215 13:00:06.941027 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 13:00:07 crc kubenswrapper[4719]: I1215 13:00:07.056847 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 13:00:07 crc kubenswrapper[4719]: I1215 13:00:07.177810 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcgw5"] Dec 15 13:00:09 crc kubenswrapper[4719]: I1215 13:00:09.033048 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gcgw5" podUID="18116782-c15c-4c02-8907-12b3a8004cf9" containerName="registry-server" containerID="cri-o://b0cf77812a3a9b6f8f4aafd49709c6eb09270fd14098e8d3752722708a8070bd" gracePeriod=2 Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.043554 4719 generic.go:334] "Generic (PLEG): container finished" podID="18116782-c15c-4c02-8907-12b3a8004cf9" containerID="b0cf77812a3a9b6f8f4aafd49709c6eb09270fd14098e8d3752722708a8070bd" exitCode=0 Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.043635 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcgw5" event={"ID":"18116782-c15c-4c02-8907-12b3a8004cf9","Type":"ContainerDied","Data":"b0cf77812a3a9b6f8f4aafd49709c6eb09270fd14098e8d3752722708a8070bd"} Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.043887 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcgw5" event={"ID":"18116782-c15c-4c02-8907-12b3a8004cf9","Type":"ContainerDied","Data":"27fcfcfb55204dd06321886b0401702e4f0932198a131c9f8250c4563eeb3f97"} Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.043910 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27fcfcfb55204dd06321886b0401702e4f0932198a131c9f8250c4563eeb3f97" Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.054117 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.175504 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-catalog-content\") pod \"18116782-c15c-4c02-8907-12b3a8004cf9\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.175595 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-utilities\") pod \"18116782-c15c-4c02-8907-12b3a8004cf9\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.175647 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99xqt\" (UniqueName: \"kubernetes.io/projected/18116782-c15c-4c02-8907-12b3a8004cf9-kube-api-access-99xqt\") pod \"18116782-c15c-4c02-8907-12b3a8004cf9\" (UID: \"18116782-c15c-4c02-8907-12b3a8004cf9\") " Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.176785 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-utilities" (OuterVolumeSpecName: "utilities") pod "18116782-c15c-4c02-8907-12b3a8004cf9" (UID: "18116782-c15c-4c02-8907-12b3a8004cf9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.184924 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18116782-c15c-4c02-8907-12b3a8004cf9-kube-api-access-99xqt" (OuterVolumeSpecName: "kube-api-access-99xqt") pod "18116782-c15c-4c02-8907-12b3a8004cf9" (UID: "18116782-c15c-4c02-8907-12b3a8004cf9"). InnerVolumeSpecName "kube-api-access-99xqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.197239 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18116782-c15c-4c02-8907-12b3a8004cf9" (UID: "18116782-c15c-4c02-8907-12b3a8004cf9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.279234 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.279388 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99xqt\" (UniqueName: \"kubernetes.io/projected/18116782-c15c-4c02-8907-12b3a8004cf9-kube-api-access-99xqt\") on node \"crc\" DevicePath \"\"" Dec 15 13:00:10 crc kubenswrapper[4719]: I1215 13:00:10.279462 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18116782-c15c-4c02-8907-12b3a8004cf9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:00:11 crc kubenswrapper[4719]: I1215 13:00:11.051350 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcgw5" Dec 15 13:00:11 crc kubenswrapper[4719]: I1215 13:00:11.085980 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcgw5"] Dec 15 13:00:11 crc kubenswrapper[4719]: I1215 13:00:11.093921 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcgw5"] Dec 15 13:00:11 crc kubenswrapper[4719]: I1215 13:00:11.494627 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18116782-c15c-4c02-8907-12b3a8004cf9" path="/var/lib/kubelet/pods/18116782-c15c-4c02-8907-12b3a8004cf9/volumes" Dec 15 13:00:17 crc kubenswrapper[4719]: I1215 13:00:17.476567 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 13:00:17 crc kubenswrapper[4719]: E1215 13:00:17.477056 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:00:29 crc kubenswrapper[4719]: I1215 13:00:29.482687 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 13:00:30 crc kubenswrapper[4719]: I1215 13:00:30.222944 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"18ecba1ded68c775cb4acec3f16c1a4fc7cc2e424616603b49749c3ef16152b6"} Dec 15 13:00:34 crc kubenswrapper[4719]: I1215 13:00:34.668947 4719 scope.go:117] "RemoveContainer" containerID="8f35255c44e421ff3bc388351b808061dae661bf0528ca8ef18a6572409ba608" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.151572 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29430061-9mfp2"] Dec 15 13:01:00 crc kubenswrapper[4719]: E1215 13:01:00.152437 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18116782-c15c-4c02-8907-12b3a8004cf9" containerName="extract-content" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.152452 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="18116782-c15c-4c02-8907-12b3a8004cf9" containerName="extract-content" Dec 15 13:01:00 crc kubenswrapper[4719]: E1215 13:01:00.152468 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29f0f318-9ef6-4f33-aec9-10cbaee5d0e7" containerName="collect-profiles" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.152474 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="29f0f318-9ef6-4f33-aec9-10cbaee5d0e7" containerName="collect-profiles" Dec 15 13:01:00 crc kubenswrapper[4719]: E1215 13:01:00.152497 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18116782-c15c-4c02-8907-12b3a8004cf9" containerName="extract-utilities" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.152504 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="18116782-c15c-4c02-8907-12b3a8004cf9" containerName="extract-utilities" Dec 15 13:01:00 crc kubenswrapper[4719]: E1215 13:01:00.152515 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18116782-c15c-4c02-8907-12b3a8004cf9" containerName="registry-server" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.152520 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="18116782-c15c-4c02-8907-12b3a8004cf9" containerName="registry-server" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.152670 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="18116782-c15c-4c02-8907-12b3a8004cf9" containerName="registry-server" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.152685 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="29f0f318-9ef6-4f33-aec9-10cbaee5d0e7" containerName="collect-profiles" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.153434 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.177237 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29430061-9mfp2"] Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.242500 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-combined-ca-bundle\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.242552 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-fernet-keys\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.242698 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-config-data\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.242792 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzwzr\" (UniqueName: \"kubernetes.io/projected/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-kube-api-access-dzwzr\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.344439 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-config-data\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.344526 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzwzr\" (UniqueName: \"kubernetes.io/projected/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-kube-api-access-dzwzr\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.344591 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-combined-ca-bundle\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.344610 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-fernet-keys\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.351081 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-fernet-keys\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.351242 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-config-data\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.352493 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-combined-ca-bundle\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.363257 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzwzr\" (UniqueName: \"kubernetes.io/projected/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-kube-api-access-dzwzr\") pod \"keystone-cron-29430061-9mfp2\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.474259 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:00 crc kubenswrapper[4719]: I1215 13:01:00.978341 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29430061-9mfp2"] Dec 15 13:01:00 crc kubenswrapper[4719]: W1215 13:01:00.982208 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32ba5a78_e7da_4c5c_9cdc_cf3e4eb18992.slice/crio-41ceebf790c2350ee09b822e925df6827e9746394521359ceb4569fd3b926a06 WatchSource:0}: Error finding container 41ceebf790c2350ee09b822e925df6827e9746394521359ceb4569fd3b926a06: Status 404 returned error can't find the container with id 41ceebf790c2350ee09b822e925df6827e9746394521359ceb4569fd3b926a06 Dec 15 13:01:01 crc kubenswrapper[4719]: I1215 13:01:01.461308 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29430061-9mfp2" event={"ID":"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992","Type":"ContainerStarted","Data":"a377b4629175e1953f839eef57677cbdc8a37f975e6e1c660fa382d914c05857"} Dec 15 13:01:01 crc kubenswrapper[4719]: I1215 13:01:01.461620 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29430061-9mfp2" event={"ID":"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992","Type":"ContainerStarted","Data":"41ceebf790c2350ee09b822e925df6827e9746394521359ceb4569fd3b926a06"} Dec 15 13:01:01 crc kubenswrapper[4719]: I1215 13:01:01.488337 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29430061-9mfp2" podStartSLOduration=1.488316492 podStartE2EDuration="1.488316492s" podCreationTimestamp="2025-12-15 13:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 13:01:01.48631969 +0000 UTC m=+2622.428612730" watchObservedRunningTime="2025-12-15 13:01:01.488316492 +0000 UTC m=+2622.430609532" Dec 15 13:01:04 crc kubenswrapper[4719]: I1215 13:01:04.488749 4719 generic.go:334] "Generic (PLEG): container finished" podID="32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992" containerID="a377b4629175e1953f839eef57677cbdc8a37f975e6e1c660fa382d914c05857" exitCode=0 Dec 15 13:01:04 crc kubenswrapper[4719]: I1215 13:01:04.488835 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29430061-9mfp2" event={"ID":"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992","Type":"ContainerDied","Data":"a377b4629175e1953f839eef57677cbdc8a37f975e6e1c660fa382d914c05857"} Dec 15 13:01:05 crc kubenswrapper[4719]: I1215 13:01:05.851112 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.003406 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-config-data\") pod \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.003493 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzwzr\" (UniqueName: \"kubernetes.io/projected/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-kube-api-access-dzwzr\") pod \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.003661 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-combined-ca-bundle\") pod \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.003760 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-fernet-keys\") pod \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\" (UID: \"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992\") " Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.011684 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992" (UID: "32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.021998 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-kube-api-access-dzwzr" (OuterVolumeSpecName: "kube-api-access-dzwzr") pod "32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992" (UID: "32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992"). InnerVolumeSpecName "kube-api-access-dzwzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.035497 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992" (UID: "32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.062618 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-config-data" (OuterVolumeSpecName: "config-data") pod "32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992" (UID: "32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.106268 4719 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.106323 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.106334 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzwzr\" (UniqueName: \"kubernetes.io/projected/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-kube-api-access-dzwzr\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.106346 4719 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.505950 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29430061-9mfp2" event={"ID":"32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992","Type":"ContainerDied","Data":"41ceebf790c2350ee09b822e925df6827e9746394521359ceb4569fd3b926a06"} Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.505989 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41ceebf790c2350ee09b822e925df6827e9746394521359ceb4569fd3b926a06" Dec 15 13:01:06 crc kubenswrapper[4719]: I1215 13:01:06.505990 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29430061-9mfp2" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.387594 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zk8dg"] Dec 15 13:01:36 crc kubenswrapper[4719]: E1215 13:01:36.388804 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992" containerName="keystone-cron" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.388824 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992" containerName="keystone-cron" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.389318 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992" containerName="keystone-cron" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.391443 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.415745 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zk8dg"] Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.508625 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-utilities\") pod \"community-operators-zk8dg\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.508938 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-catalog-content\") pod \"community-operators-zk8dg\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.509005 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxkjn\" (UniqueName: \"kubernetes.io/projected/c2033e17-b2d7-48f7-a711-58f63902069d-kube-api-access-wxkjn\") pod \"community-operators-zk8dg\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.610338 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-catalog-content\") pod \"community-operators-zk8dg\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.610429 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxkjn\" (UniqueName: \"kubernetes.io/projected/c2033e17-b2d7-48f7-a711-58f63902069d-kube-api-access-wxkjn\") pod \"community-operators-zk8dg\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.610483 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-utilities\") pod \"community-operators-zk8dg\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.610922 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-catalog-content\") pod \"community-operators-zk8dg\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.610959 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-utilities\") pod \"community-operators-zk8dg\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.639096 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxkjn\" (UniqueName: \"kubernetes.io/projected/c2033e17-b2d7-48f7-a711-58f63902069d-kube-api-access-wxkjn\") pod \"community-operators-zk8dg\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:36 crc kubenswrapper[4719]: I1215 13:01:36.719729 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:37 crc kubenswrapper[4719]: I1215 13:01:37.314494 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zk8dg"] Dec 15 13:01:37 crc kubenswrapper[4719]: I1215 13:01:37.768770 4719 generic.go:334] "Generic (PLEG): container finished" podID="c2033e17-b2d7-48f7-a711-58f63902069d" containerID="641deb65b09e4b8a13c7c1d5eea6aa4076bb554de45c690374870e899d97a543" exitCode=0 Dec 15 13:01:37 crc kubenswrapper[4719]: I1215 13:01:37.768884 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk8dg" event={"ID":"c2033e17-b2d7-48f7-a711-58f63902069d","Type":"ContainerDied","Data":"641deb65b09e4b8a13c7c1d5eea6aa4076bb554de45c690374870e899d97a543"} Dec 15 13:01:37 crc kubenswrapper[4719]: I1215 13:01:37.769177 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk8dg" event={"ID":"c2033e17-b2d7-48f7-a711-58f63902069d","Type":"ContainerStarted","Data":"6f04112edc3bf33b080f0ea8b7582793a2a7bc3bd484e3f19fd93c5a15796885"} Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.209321 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j7xqn"] Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.211924 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.226695 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j7xqn"] Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.344847 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aed14ad-077e-449b-a183-d583dc39893d-utilities\") pod \"redhat-operators-j7xqn\" (UID: \"6aed14ad-077e-449b-a183-d583dc39893d\") " pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.344935 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltx47\" (UniqueName: \"kubernetes.io/projected/6aed14ad-077e-449b-a183-d583dc39893d-kube-api-access-ltx47\") pod \"redhat-operators-j7xqn\" (UID: \"6aed14ad-077e-449b-a183-d583dc39893d\") " pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.345067 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aed14ad-077e-449b-a183-d583dc39893d-catalog-content\") pod \"redhat-operators-j7xqn\" (UID: \"6aed14ad-077e-449b-a183-d583dc39893d\") " pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.446982 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aed14ad-077e-449b-a183-d583dc39893d-catalog-content\") pod \"redhat-operators-j7xqn\" (UID: \"6aed14ad-077e-449b-a183-d583dc39893d\") " pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.447049 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aed14ad-077e-449b-a183-d583dc39893d-utilities\") pod \"redhat-operators-j7xqn\" (UID: \"6aed14ad-077e-449b-a183-d583dc39893d\") " pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.447091 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltx47\" (UniqueName: \"kubernetes.io/projected/6aed14ad-077e-449b-a183-d583dc39893d-kube-api-access-ltx47\") pod \"redhat-operators-j7xqn\" (UID: \"6aed14ad-077e-449b-a183-d583dc39893d\") " pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.447731 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aed14ad-077e-449b-a183-d583dc39893d-catalog-content\") pod \"redhat-operators-j7xqn\" (UID: \"6aed14ad-077e-449b-a183-d583dc39893d\") " pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.447765 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aed14ad-077e-449b-a183-d583dc39893d-utilities\") pod \"redhat-operators-j7xqn\" (UID: \"6aed14ad-077e-449b-a183-d583dc39893d\") " pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.466619 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltx47\" (UniqueName: \"kubernetes.io/projected/6aed14ad-077e-449b-a183-d583dc39893d-kube-api-access-ltx47\") pod \"redhat-operators-j7xqn\" (UID: \"6aed14ad-077e-449b-a183-d583dc39893d\") " pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:38 crc kubenswrapper[4719]: I1215 13:01:38.547449 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.030443 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j7xqn"] Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.580334 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mptcj"] Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.582627 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.596272 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mptcj"] Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.775526 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p49j5\" (UniqueName: \"kubernetes.io/projected/ee0096c1-a669-470d-ac7e-cab99afd927a-kube-api-access-p49j5\") pod \"certified-operators-mptcj\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.775594 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-utilities\") pod \"certified-operators-mptcj\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.775691 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-catalog-content\") pod \"certified-operators-mptcj\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.790516 4719 generic.go:334] "Generic (PLEG): container finished" podID="c2033e17-b2d7-48f7-a711-58f63902069d" containerID="45bca6d816ceb05f6ed2567cbd694e37047cad386f5255fcbe24b38d11cd6e2c" exitCode=0 Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.790763 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk8dg" event={"ID":"c2033e17-b2d7-48f7-a711-58f63902069d","Type":"ContainerDied","Data":"45bca6d816ceb05f6ed2567cbd694e37047cad386f5255fcbe24b38d11cd6e2c"} Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.794027 4719 generic.go:334] "Generic (PLEG): container finished" podID="6aed14ad-077e-449b-a183-d583dc39893d" containerID="287d64303db887d60598c6a0697871378cda2838766ab5ad8ee26646d1052af7" exitCode=0 Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.794068 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7xqn" event={"ID":"6aed14ad-077e-449b-a183-d583dc39893d","Type":"ContainerDied","Data":"287d64303db887d60598c6a0697871378cda2838766ab5ad8ee26646d1052af7"} Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.794096 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7xqn" event={"ID":"6aed14ad-077e-449b-a183-d583dc39893d","Type":"ContainerStarted","Data":"4c6dd6fb43f07b69a8656bb8d860e98bdc09c6942ae125cf33d3f959518c030c"} Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.877269 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-catalog-content\") pod \"certified-operators-mptcj\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.877386 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p49j5\" (UniqueName: \"kubernetes.io/projected/ee0096c1-a669-470d-ac7e-cab99afd927a-kube-api-access-p49j5\") pod \"certified-operators-mptcj\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.877417 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-utilities\") pod \"certified-operators-mptcj\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.877801 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-catalog-content\") pod \"certified-operators-mptcj\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.877921 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-utilities\") pod \"certified-operators-mptcj\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:39 crc kubenswrapper[4719]: I1215 13:01:39.908707 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p49j5\" (UniqueName: \"kubernetes.io/projected/ee0096c1-a669-470d-ac7e-cab99afd927a-kube-api-access-p49j5\") pod \"certified-operators-mptcj\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:40 crc kubenswrapper[4719]: I1215 13:01:40.201922 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:01:40 crc kubenswrapper[4719]: I1215 13:01:40.792007 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mptcj"] Dec 15 13:01:40 crc kubenswrapper[4719]: I1215 13:01:40.810102 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mptcj" event={"ID":"ee0096c1-a669-470d-ac7e-cab99afd927a","Type":"ContainerStarted","Data":"8dd03f2e8d5b705fd025d6e37174e48d3c7de60baef44854c80e32c86be8c2c4"} Dec 15 13:01:41 crc kubenswrapper[4719]: I1215 13:01:41.837277 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mptcj" event={"ID":"ee0096c1-a669-470d-ac7e-cab99afd927a","Type":"ContainerStarted","Data":"5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd"} Dec 15 13:01:41 crc kubenswrapper[4719]: I1215 13:01:41.842838 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk8dg" event={"ID":"c2033e17-b2d7-48f7-a711-58f63902069d","Type":"ContainerStarted","Data":"b2a08e4b1890d64b3b8201ce3fe315bddc93b19d8abea31b6d0028fdb21eb67d"} Dec 15 13:01:41 crc kubenswrapper[4719]: I1215 13:01:41.870748 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zk8dg" podStartSLOduration=3.202833018 podStartE2EDuration="5.870720564s" podCreationTimestamp="2025-12-15 13:01:36 +0000 UTC" firstStartedPulling="2025-12-15 13:01:37.770417267 +0000 UTC m=+2658.712710307" lastFinishedPulling="2025-12-15 13:01:40.438304823 +0000 UTC m=+2661.380597853" observedRunningTime="2025-12-15 13:01:41.862001424 +0000 UTC m=+2662.804294454" watchObservedRunningTime="2025-12-15 13:01:41.870720564 +0000 UTC m=+2662.813013594" Dec 15 13:01:43 crc kubenswrapper[4719]: I1215 13:01:43.864660 4719 generic.go:334] "Generic (PLEG): container finished" podID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerID="5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd" exitCode=0 Dec 15 13:01:43 crc kubenswrapper[4719]: I1215 13:01:43.864698 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mptcj" event={"ID":"ee0096c1-a669-470d-ac7e-cab99afd927a","Type":"ContainerDied","Data":"5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd"} Dec 15 13:01:44 crc kubenswrapper[4719]: I1215 13:01:44.876674 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mptcj" event={"ID":"ee0096c1-a669-470d-ac7e-cab99afd927a","Type":"ContainerStarted","Data":"0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b"} Dec 15 13:01:46 crc kubenswrapper[4719]: I1215 13:01:46.720716 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:46 crc kubenswrapper[4719]: I1215 13:01:46.721044 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:46 crc kubenswrapper[4719]: I1215 13:01:46.772415 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:46 crc kubenswrapper[4719]: I1215 13:01:46.896666 4719 generic.go:334] "Generic (PLEG): container finished" podID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerID="0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b" exitCode=0 Dec 15 13:01:46 crc kubenswrapper[4719]: I1215 13:01:46.896849 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mptcj" event={"ID":"ee0096c1-a669-470d-ac7e-cab99afd927a","Type":"ContainerDied","Data":"0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b"} Dec 15 13:01:46 crc kubenswrapper[4719]: I1215 13:01:46.946812 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:49 crc kubenswrapper[4719]: I1215 13:01:49.180000 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zk8dg"] Dec 15 13:01:49 crc kubenswrapper[4719]: I1215 13:01:49.182204 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zk8dg" podUID="c2033e17-b2d7-48f7-a711-58f63902069d" containerName="registry-server" containerID="cri-o://b2a08e4b1890d64b3b8201ce3fe315bddc93b19d8abea31b6d0028fdb21eb67d" gracePeriod=2 Dec 15 13:01:50 crc kubenswrapper[4719]: I1215 13:01:50.934968 4719 generic.go:334] "Generic (PLEG): container finished" podID="c2033e17-b2d7-48f7-a711-58f63902069d" containerID="b2a08e4b1890d64b3b8201ce3fe315bddc93b19d8abea31b6d0028fdb21eb67d" exitCode=0 Dec 15 13:01:50 crc kubenswrapper[4719]: I1215 13:01:50.935066 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk8dg" event={"ID":"c2033e17-b2d7-48f7-a711-58f63902069d","Type":"ContainerDied","Data":"b2a08e4b1890d64b3b8201ce3fe315bddc93b19d8abea31b6d0028fdb21eb67d"} Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.196010 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.385622 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-utilities\") pod \"c2033e17-b2d7-48f7-a711-58f63902069d\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.385959 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkjn\" (UniqueName: \"kubernetes.io/projected/c2033e17-b2d7-48f7-a711-58f63902069d-kube-api-access-wxkjn\") pod \"c2033e17-b2d7-48f7-a711-58f63902069d\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.386183 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-catalog-content\") pod \"c2033e17-b2d7-48f7-a711-58f63902069d\" (UID: \"c2033e17-b2d7-48f7-a711-58f63902069d\") " Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.388359 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-utilities" (OuterVolumeSpecName: "utilities") pod "c2033e17-b2d7-48f7-a711-58f63902069d" (UID: "c2033e17-b2d7-48f7-a711-58f63902069d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.395143 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2033e17-b2d7-48f7-a711-58f63902069d-kube-api-access-wxkjn" (OuterVolumeSpecName: "kube-api-access-wxkjn") pod "c2033e17-b2d7-48f7-a711-58f63902069d" (UID: "c2033e17-b2d7-48f7-a711-58f63902069d"). InnerVolumeSpecName "kube-api-access-wxkjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.438960 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2033e17-b2d7-48f7-a711-58f63902069d" (UID: "c2033e17-b2d7-48f7-a711-58f63902069d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.488687 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.489016 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2033e17-b2d7-48f7-a711-58f63902069d-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.489081 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkjn\" (UniqueName: \"kubernetes.io/projected/c2033e17-b2d7-48f7-a711-58f63902069d-kube-api-access-wxkjn\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.959703 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7xqn" event={"ID":"6aed14ad-077e-449b-a183-d583dc39893d","Type":"ContainerStarted","Data":"a4e09b53c54849277a2e4f98a6810b319787a4b6aaf1a590a10e8f7c57fc7bb8"} Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.963548 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk8dg" event={"ID":"c2033e17-b2d7-48f7-a711-58f63902069d","Type":"ContainerDied","Data":"6f04112edc3bf33b080f0ea8b7582793a2a7bc3bd484e3f19fd93c5a15796885"} Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.963587 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk8dg" Dec 15 13:01:52 crc kubenswrapper[4719]: I1215 13:01:52.963593 4719 scope.go:117] "RemoveContainer" containerID="b2a08e4b1890d64b3b8201ce3fe315bddc93b19d8abea31b6d0028fdb21eb67d" Dec 15 13:01:53 crc kubenswrapper[4719]: I1215 13:01:53.013887 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zk8dg"] Dec 15 13:01:53 crc kubenswrapper[4719]: I1215 13:01:53.023439 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zk8dg"] Dec 15 13:01:53 crc kubenswrapper[4719]: I1215 13:01:53.136655 4719 scope.go:117] "RemoveContainer" containerID="45bca6d816ceb05f6ed2567cbd694e37047cad386f5255fcbe24b38d11cd6e2c" Dec 15 13:01:53 crc kubenswrapper[4719]: I1215 13:01:53.158737 4719 scope.go:117] "RemoveContainer" containerID="641deb65b09e4b8a13c7c1d5eea6aa4076bb554de45c690374870e899d97a543" Dec 15 13:01:53 crc kubenswrapper[4719]: I1215 13:01:53.513695 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2033e17-b2d7-48f7-a711-58f63902069d" path="/var/lib/kubelet/pods/c2033e17-b2d7-48f7-a711-58f63902069d/volumes" Dec 15 13:01:53 crc kubenswrapper[4719]: I1215 13:01:53.974416 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mptcj" event={"ID":"ee0096c1-a669-470d-ac7e-cab99afd927a","Type":"ContainerStarted","Data":"1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5"} Dec 15 13:01:54 crc kubenswrapper[4719]: I1215 13:01:54.000284 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mptcj" podStartSLOduration=5.241542702 podStartE2EDuration="15.000266318s" podCreationTimestamp="2025-12-15 13:01:39 +0000 UTC" firstStartedPulling="2025-12-15 13:01:43.868065118 +0000 UTC m=+2664.810358148" lastFinishedPulling="2025-12-15 13:01:53.626788724 +0000 UTC m=+2674.569081764" observedRunningTime="2025-12-15 13:01:53.993710405 +0000 UTC m=+2674.936003445" watchObservedRunningTime="2025-12-15 13:01:54.000266318 +0000 UTC m=+2674.942559348" Dec 15 13:01:55 crc kubenswrapper[4719]: I1215 13:01:55.991395 4719 generic.go:334] "Generic (PLEG): container finished" podID="6aed14ad-077e-449b-a183-d583dc39893d" containerID="a4e09b53c54849277a2e4f98a6810b319787a4b6aaf1a590a10e8f7c57fc7bb8" exitCode=0 Dec 15 13:01:55 crc kubenswrapper[4719]: I1215 13:01:55.991475 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7xqn" event={"ID":"6aed14ad-077e-449b-a183-d583dc39893d","Type":"ContainerDied","Data":"a4e09b53c54849277a2e4f98a6810b319787a4b6aaf1a590a10e8f7c57fc7bb8"} Dec 15 13:01:57 crc kubenswrapper[4719]: I1215 13:01:57.000785 4719 generic.go:334] "Generic (PLEG): container finished" podID="c390d0c5-24b7-4810-baf1-d7b9fd9dd13f" containerID="b28f0e9b50803bf3e3722620e6dc2fc9fb15836f3a4f4fb9f9a1a4189a7ef625" exitCode=0 Dec 15 13:01:57 crc kubenswrapper[4719]: I1215 13:01:57.000933 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" event={"ID":"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f","Type":"ContainerDied","Data":"b28f0e9b50803bf3e3722620e6dc2fc9fb15836f3a4f4fb9f9a1a4189a7ef625"} Dec 15 13:01:57 crc kubenswrapper[4719]: I1215 13:01:57.004582 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j7xqn" event={"ID":"6aed14ad-077e-449b-a183-d583dc39893d","Type":"ContainerStarted","Data":"a261eef617843526b097e8ae0bb55828e646d0eadcfe8035deef8387f10bdf2c"} Dec 15 13:01:57 crc kubenswrapper[4719]: I1215 13:01:57.038582 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j7xqn" podStartSLOduration=2.242679535 podStartE2EDuration="19.038560742s" podCreationTimestamp="2025-12-15 13:01:38 +0000 UTC" firstStartedPulling="2025-12-15 13:01:39.796310954 +0000 UTC m=+2660.738603994" lastFinishedPulling="2025-12-15 13:01:56.592192171 +0000 UTC m=+2677.534485201" observedRunningTime="2025-12-15 13:01:57.036373484 +0000 UTC m=+2677.978666514" watchObservedRunningTime="2025-12-15 13:01:57.038560742 +0000 UTC m=+2677.980853782" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.448240 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.521481 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-ssh-key\") pod \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.521606 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-inventory\") pod \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.521652 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-secret-0\") pod \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.521688 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-combined-ca-bundle\") pod \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.521715 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crv59\" (UniqueName: \"kubernetes.io/projected/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-kube-api-access-crv59\") pod \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\" (UID: \"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f\") " Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.526978 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c390d0c5-24b7-4810-baf1-d7b9fd9dd13f" (UID: "c390d0c5-24b7-4810-baf1-d7b9fd9dd13f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.527370 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-kube-api-access-crv59" (OuterVolumeSpecName: "kube-api-access-crv59") pod "c390d0c5-24b7-4810-baf1-d7b9fd9dd13f" (UID: "c390d0c5-24b7-4810-baf1-d7b9fd9dd13f"). InnerVolumeSpecName "kube-api-access-crv59". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.548646 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.548721 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.551626 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-inventory" (OuterVolumeSpecName: "inventory") pod "c390d0c5-24b7-4810-baf1-d7b9fd9dd13f" (UID: "c390d0c5-24b7-4810-baf1-d7b9fd9dd13f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.561648 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "c390d0c5-24b7-4810-baf1-d7b9fd9dd13f" (UID: "c390d0c5-24b7-4810-baf1-d7b9fd9dd13f"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.573167 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c390d0c5-24b7-4810-baf1-d7b9fd9dd13f" (UID: "c390d0c5-24b7-4810-baf1-d7b9fd9dd13f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.623835 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.623889 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.623899 4719 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.623908 4719 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:58 crc kubenswrapper[4719]: I1215 13:01:58.623919 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crv59\" (UniqueName: \"kubernetes.io/projected/c390d0c5-24b7-4810-baf1-d7b9fd9dd13f-kube-api-access-crv59\") on node \"crc\" DevicePath \"\"" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.036494 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.037607 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql" event={"ID":"c390d0c5-24b7-4810-baf1-d7b9fd9dd13f","Type":"ContainerDied","Data":"b0a5893a3e1deaf712855224fd680d4368777a4f8132235738611c1a8aa3e20b"} Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.037649 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0a5893a3e1deaf712855224fd680d4368777a4f8132235738611c1a8aa3e20b" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.145142 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9"] Dec 15 13:01:59 crc kubenswrapper[4719]: E1215 13:01:59.145717 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2033e17-b2d7-48f7-a711-58f63902069d" containerName="extract-content" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.145746 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2033e17-b2d7-48f7-a711-58f63902069d" containerName="extract-content" Dec 15 13:01:59 crc kubenswrapper[4719]: E1215 13:01:59.145760 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2033e17-b2d7-48f7-a711-58f63902069d" containerName="extract-utilities" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.145769 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2033e17-b2d7-48f7-a711-58f63902069d" containerName="extract-utilities" Dec 15 13:01:59 crc kubenswrapper[4719]: E1215 13:01:59.145807 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c390d0c5-24b7-4810-baf1-d7b9fd9dd13f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.145816 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c390d0c5-24b7-4810-baf1-d7b9fd9dd13f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 15 13:01:59 crc kubenswrapper[4719]: E1215 13:01:59.145835 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2033e17-b2d7-48f7-a711-58f63902069d" containerName="registry-server" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.145843 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2033e17-b2d7-48f7-a711-58f63902069d" containerName="registry-server" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.146099 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c390d0c5-24b7-4810-baf1-d7b9fd9dd13f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.146121 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2033e17-b2d7-48f7-a711-58f63902069d" containerName="registry-server" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.146945 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.156874 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9"] Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.157470 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.157707 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.157881 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.158025 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.158094 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.158690 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.159776 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.334350 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.334719 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.334791 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.334828 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.334849 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.334899 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.334992 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.335073 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmm8q\" (UniqueName: \"kubernetes.io/projected/48c18077-569a-4bbd-ace2-64dbf06bcbee-kube-api-access-jmm8q\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.335107 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.436292 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.436559 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.436582 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.436601 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.436627 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.436654 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.436724 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmm8q\" (UniqueName: \"kubernetes.io/projected/48c18077-569a-4bbd-ace2-64dbf06bcbee-kube-api-access-jmm8q\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.436751 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.436771 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.438871 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.440628 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.441168 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.441555 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.441810 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.442400 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.443286 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.451639 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.454467 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmm8q\" (UniqueName: \"kubernetes.io/projected/48c18077-569a-4bbd-ace2-64dbf06bcbee-kube-api-access-jmm8q\") pod \"nova-edpm-deployment-openstack-edpm-ipam-hvqz9\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.464640 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:01:59 crc kubenswrapper[4719]: I1215 13:01:59.601201 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j7xqn" podUID="6aed14ad-077e-449b-a183-d583dc39893d" containerName="registry-server" probeResult="failure" output=< Dec 15 13:01:59 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 13:01:59 crc kubenswrapper[4719]: > Dec 15 13:02:00 crc kubenswrapper[4719]: I1215 13:02:00.005620 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9"] Dec 15 13:02:00 crc kubenswrapper[4719]: I1215 13:02:00.054869 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" event={"ID":"48c18077-569a-4bbd-ace2-64dbf06bcbee","Type":"ContainerStarted","Data":"11484a88a77244482cff5caabe7639745a58c61bda25930618db3feca8db34f1"} Dec 15 13:02:00 crc kubenswrapper[4719]: I1215 13:02:00.203080 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:02:00 crc kubenswrapper[4719]: I1215 13:02:00.203144 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:02:01 crc kubenswrapper[4719]: I1215 13:02:01.064462 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" event={"ID":"48c18077-569a-4bbd-ace2-64dbf06bcbee","Type":"ContainerStarted","Data":"d24a3aa5f6c9740f53d3b20412f2f3d963ba2bd2d8b692f8f325205731660e94"} Dec 15 13:02:01 crc kubenswrapper[4719]: I1215 13:02:01.089878 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" podStartSLOduration=1.472300289 podStartE2EDuration="2.089843381s" podCreationTimestamp="2025-12-15 13:01:59 +0000 UTC" firstStartedPulling="2025-12-15 13:02:00.034580187 +0000 UTC m=+2680.976873217" lastFinishedPulling="2025-12-15 13:02:00.652123279 +0000 UTC m=+2681.594416309" observedRunningTime="2025-12-15 13:02:01.082788813 +0000 UTC m=+2682.025081853" watchObservedRunningTime="2025-12-15 13:02:01.089843381 +0000 UTC m=+2682.032136411" Dec 15 13:02:01 crc kubenswrapper[4719]: I1215 13:02:01.250697 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mptcj" podUID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerName="registry-server" probeResult="failure" output=< Dec 15 13:02:01 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 13:02:01 crc kubenswrapper[4719]: > Dec 15 13:02:08 crc kubenswrapper[4719]: I1215 13:02:08.630985 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:02:08 crc kubenswrapper[4719]: I1215 13:02:08.706173 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j7xqn" Dec 15 13:02:09 crc kubenswrapper[4719]: I1215 13:02:09.239173 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j7xqn"] Dec 15 13:02:09 crc kubenswrapper[4719]: I1215 13:02:09.390768 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6qhp2"] Dec 15 13:02:09 crc kubenswrapper[4719]: I1215 13:02:09.391041 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6qhp2" podUID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerName="registry-server" containerID="cri-o://0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd" gracePeriod=2 Dec 15 13:02:09 crc kubenswrapper[4719]: I1215 13:02:09.938722 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.070494 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-utilities\") pod \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.070608 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-catalog-content\") pod \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.070754 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqsnr\" (UniqueName: \"kubernetes.io/projected/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-kube-api-access-qqsnr\") pod \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\" (UID: \"56dfcdfa-9677-498d-8f13-eaf8cabc48e3\") " Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.071626 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-utilities" (OuterVolumeSpecName: "utilities") pod "56dfcdfa-9677-498d-8f13-eaf8cabc48e3" (UID: "56dfcdfa-9677-498d-8f13-eaf8cabc48e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.087837 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-kube-api-access-qqsnr" (OuterVolumeSpecName: "kube-api-access-qqsnr") pod "56dfcdfa-9677-498d-8f13-eaf8cabc48e3" (UID: "56dfcdfa-9677-498d-8f13-eaf8cabc48e3"). InnerVolumeSpecName "kube-api-access-qqsnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.167470 4719 generic.go:334] "Generic (PLEG): container finished" podID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerID="0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd" exitCode=0 Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.167531 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qhp2" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.167625 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qhp2" event={"ID":"56dfcdfa-9677-498d-8f13-eaf8cabc48e3","Type":"ContainerDied","Data":"0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd"} Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.167656 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qhp2" event={"ID":"56dfcdfa-9677-498d-8f13-eaf8cabc48e3","Type":"ContainerDied","Data":"08d2af0a83ec4eb7d4a0254c84ecab632f260b7e768243862958fd64a90fe71a"} Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.167671 4719 scope.go:117] "RemoveContainer" containerID="0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.172945 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.173001 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqsnr\" (UniqueName: \"kubernetes.io/projected/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-kube-api-access-qqsnr\") on node \"crc\" DevicePath \"\"" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.186343 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56dfcdfa-9677-498d-8f13-eaf8cabc48e3" (UID: "56dfcdfa-9677-498d-8f13-eaf8cabc48e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.205637 4719 scope.go:117] "RemoveContainer" containerID="5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.247103 4719 scope.go:117] "RemoveContainer" containerID="55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.268537 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.274899 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56dfcdfa-9677-498d-8f13-eaf8cabc48e3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.303140 4719 scope.go:117] "RemoveContainer" containerID="0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd" Dec 15 13:02:10 crc kubenswrapper[4719]: E1215 13:02:10.304149 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd\": container with ID starting with 0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd not found: ID does not exist" containerID="0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.304187 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd"} err="failed to get container status \"0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd\": rpc error: code = NotFound desc = could not find container \"0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd\": container with ID starting with 0c69aa27859d682c61853b73b605ed6477ba226ecbfeeaf362c0da96dfae42bd not found: ID does not exist" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.304207 4719 scope.go:117] "RemoveContainer" containerID="5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4" Dec 15 13:02:10 crc kubenswrapper[4719]: E1215 13:02:10.304552 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4\": container with ID starting with 5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4 not found: ID does not exist" containerID="5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.304589 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4"} err="failed to get container status \"5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4\": rpc error: code = NotFound desc = could not find container \"5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4\": container with ID starting with 5d28349c7da3882f278e2d2f9b96ae29cbd16317a411118155dddb88f48447c4 not found: ID does not exist" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.304602 4719 scope.go:117] "RemoveContainer" containerID="55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619" Dec 15 13:02:10 crc kubenswrapper[4719]: E1215 13:02:10.304794 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619\": container with ID starting with 55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619 not found: ID does not exist" containerID="55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.304914 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619"} err="failed to get container status \"55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619\": rpc error: code = NotFound desc = could not find container \"55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619\": container with ID starting with 55099cb34f6fb50f96b2ed4b9109736d09c832cfe2e4e4db182bb1922e355619 not found: ID does not exist" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.324835 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.498507 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6qhp2"] Dec 15 13:02:10 crc kubenswrapper[4719]: I1215 13:02:10.508962 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6qhp2"] Dec 15 13:02:11 crc kubenswrapper[4719]: I1215 13:02:11.486323 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" path="/var/lib/kubelet/pods/56dfcdfa-9677-498d-8f13-eaf8cabc48e3/volumes" Dec 15 13:02:12 crc kubenswrapper[4719]: I1215 13:02:12.601151 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mptcj"] Dec 15 13:02:12 crc kubenswrapper[4719]: I1215 13:02:12.601840 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mptcj" podUID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerName="registry-server" containerID="cri-o://1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5" gracePeriod=2 Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.061225 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.126688 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-catalog-content\") pod \"ee0096c1-a669-470d-ac7e-cab99afd927a\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.126804 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-utilities\") pod \"ee0096c1-a669-470d-ac7e-cab99afd927a\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.126906 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p49j5\" (UniqueName: \"kubernetes.io/projected/ee0096c1-a669-470d-ac7e-cab99afd927a-kube-api-access-p49j5\") pod \"ee0096c1-a669-470d-ac7e-cab99afd927a\" (UID: \"ee0096c1-a669-470d-ac7e-cab99afd927a\") " Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.127330 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-utilities" (OuterVolumeSpecName: "utilities") pod "ee0096c1-a669-470d-ac7e-cab99afd927a" (UID: "ee0096c1-a669-470d-ac7e-cab99afd927a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.132318 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee0096c1-a669-470d-ac7e-cab99afd927a-kube-api-access-p49j5" (OuterVolumeSpecName: "kube-api-access-p49j5") pod "ee0096c1-a669-470d-ac7e-cab99afd927a" (UID: "ee0096c1-a669-470d-ac7e-cab99afd927a"). InnerVolumeSpecName "kube-api-access-p49j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.189136 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee0096c1-a669-470d-ac7e-cab99afd927a" (UID: "ee0096c1-a669-470d-ac7e-cab99afd927a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.197339 4719 generic.go:334] "Generic (PLEG): container finished" podID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerID="1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5" exitCode=0 Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.197373 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mptcj" event={"ID":"ee0096c1-a669-470d-ac7e-cab99afd927a","Type":"ContainerDied","Data":"1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5"} Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.197396 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mptcj" event={"ID":"ee0096c1-a669-470d-ac7e-cab99afd927a","Type":"ContainerDied","Data":"8dd03f2e8d5b705fd025d6e37174e48d3c7de60baef44854c80e32c86be8c2c4"} Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.197414 4719 scope.go:117] "RemoveContainer" containerID="1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.197498 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mptcj" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.230670 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.230718 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p49j5\" (UniqueName: \"kubernetes.io/projected/ee0096c1-a669-470d-ac7e-cab99afd927a-kube-api-access-p49j5\") on node \"crc\" DevicePath \"\"" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.235158 4719 scope.go:117] "RemoveContainer" containerID="0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.235234 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee0096c1-a669-470d-ac7e-cab99afd927a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.264638 4719 scope.go:117] "RemoveContainer" containerID="5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.279325 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mptcj"] Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.293308 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mptcj"] Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.315155 4719 scope.go:117] "RemoveContainer" containerID="1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5" Dec 15 13:02:13 crc kubenswrapper[4719]: E1215 13:02:13.315465 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5\": container with ID starting with 1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5 not found: ID does not exist" containerID="1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.315490 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5"} err="failed to get container status \"1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5\": rpc error: code = NotFound desc = could not find container \"1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5\": container with ID starting with 1ddfbbe0423fdb148331cf9bdc11828aa167a5864000b0208de7b78df809e9e5 not found: ID does not exist" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.315511 4719 scope.go:117] "RemoveContainer" containerID="0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b" Dec 15 13:02:13 crc kubenswrapper[4719]: E1215 13:02:13.315668 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b\": container with ID starting with 0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b not found: ID does not exist" containerID="0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.315684 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b"} err="failed to get container status \"0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b\": rpc error: code = NotFound desc = could not find container \"0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b\": container with ID starting with 0125d0d20e66ccef0314d60924572af695a9028da85d110efd0ddb14854aec3b not found: ID does not exist" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.315695 4719 scope.go:117] "RemoveContainer" containerID="5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd" Dec 15 13:02:13 crc kubenswrapper[4719]: E1215 13:02:13.315868 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd\": container with ID starting with 5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd not found: ID does not exist" containerID="5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.315883 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd"} err="failed to get container status \"5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd\": rpc error: code = NotFound desc = could not find container \"5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd\": container with ID starting with 5239e218f8d7aeab12488fb45694cf80d6d0c89d868d246e31402c895d1c3cdd not found: ID does not exist" Dec 15 13:02:13 crc kubenswrapper[4719]: I1215 13:02:13.484398 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee0096c1-a669-470d-ac7e-cab99afd927a" path="/var/lib/kubelet/pods/ee0096c1-a669-470d-ac7e-cab99afd927a/volumes" Dec 15 13:02:51 crc kubenswrapper[4719]: I1215 13:02:51.356471 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:02:51 crc kubenswrapper[4719]: I1215 13:02:51.357037 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:03:21 crc kubenswrapper[4719]: I1215 13:03:21.357426 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:03:21 crc kubenswrapper[4719]: I1215 13:03:21.357887 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:03:51 crc kubenswrapper[4719]: I1215 13:03:51.357195 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:03:51 crc kubenswrapper[4719]: I1215 13:03:51.357825 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:03:51 crc kubenswrapper[4719]: I1215 13:03:51.357939 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 13:03:51 crc kubenswrapper[4719]: I1215 13:03:51.358638 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"18ecba1ded68c775cb4acec3f16c1a4fc7cc2e424616603b49749c3ef16152b6"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 13:03:51 crc kubenswrapper[4719]: I1215 13:03:51.358693 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://18ecba1ded68c775cb4acec3f16c1a4fc7cc2e424616603b49749c3ef16152b6" gracePeriod=600 Dec 15 13:03:52 crc kubenswrapper[4719]: I1215 13:03:52.047645 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="18ecba1ded68c775cb4acec3f16c1a4fc7cc2e424616603b49749c3ef16152b6" exitCode=0 Dec 15 13:03:52 crc kubenswrapper[4719]: I1215 13:03:52.047718 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"18ecba1ded68c775cb4acec3f16c1a4fc7cc2e424616603b49749c3ef16152b6"} Dec 15 13:03:52 crc kubenswrapper[4719]: I1215 13:03:52.048129 4719 scope.go:117] "RemoveContainer" containerID="50dfc78db912eaf77ae556050df439cd24cba96cbfe5e62eb5306b7a51b1ba58" Dec 15 13:03:53 crc kubenswrapper[4719]: I1215 13:03:53.056875 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d"} Dec 15 13:04:54 crc kubenswrapper[4719]: I1215 13:04:54.610798 4719 generic.go:334] "Generic (PLEG): container finished" podID="48c18077-569a-4bbd-ace2-64dbf06bcbee" containerID="d24a3aa5f6c9740f53d3b20412f2f3d963ba2bd2d8b692f8f325205731660e94" exitCode=0 Dec 15 13:04:54 crc kubenswrapper[4719]: I1215 13:04:54.610883 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" event={"ID":"48c18077-569a-4bbd-ace2-64dbf06bcbee","Type":"ContainerDied","Data":"d24a3aa5f6c9740f53d3b20412f2f3d963ba2bd2d8b692f8f325205731660e94"} Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.067664 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.135930 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-0\") pod \"48c18077-569a-4bbd-ace2-64dbf06bcbee\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.136023 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-combined-ca-bundle\") pod \"48c18077-569a-4bbd-ace2-64dbf06bcbee\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.136073 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-extra-config-0\") pod \"48c18077-569a-4bbd-ace2-64dbf06bcbee\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.136101 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-inventory\") pod \"48c18077-569a-4bbd-ace2-64dbf06bcbee\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.136124 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-1\") pod \"48c18077-569a-4bbd-ace2-64dbf06bcbee\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.136181 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-ssh-key\") pod \"48c18077-569a-4bbd-ace2-64dbf06bcbee\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.136307 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmm8q\" (UniqueName: \"kubernetes.io/projected/48c18077-569a-4bbd-ace2-64dbf06bcbee-kube-api-access-jmm8q\") pod \"48c18077-569a-4bbd-ace2-64dbf06bcbee\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.136348 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-0\") pod \"48c18077-569a-4bbd-ace2-64dbf06bcbee\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.136396 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-1\") pod \"48c18077-569a-4bbd-ace2-64dbf06bcbee\" (UID: \"48c18077-569a-4bbd-ace2-64dbf06bcbee\") " Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.147085 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "48c18077-569a-4bbd-ace2-64dbf06bcbee" (UID: "48c18077-569a-4bbd-ace2-64dbf06bcbee"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.149278 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48c18077-569a-4bbd-ace2-64dbf06bcbee-kube-api-access-jmm8q" (OuterVolumeSpecName: "kube-api-access-jmm8q") pod "48c18077-569a-4bbd-ace2-64dbf06bcbee" (UID: "48c18077-569a-4bbd-ace2-64dbf06bcbee"). InnerVolumeSpecName "kube-api-access-jmm8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.165555 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "48c18077-569a-4bbd-ace2-64dbf06bcbee" (UID: "48c18077-569a-4bbd-ace2-64dbf06bcbee"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.175487 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "48c18077-569a-4bbd-ace2-64dbf06bcbee" (UID: "48c18077-569a-4bbd-ace2-64dbf06bcbee"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.177733 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "48c18077-569a-4bbd-ace2-64dbf06bcbee" (UID: "48c18077-569a-4bbd-ace2-64dbf06bcbee"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.180636 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "48c18077-569a-4bbd-ace2-64dbf06bcbee" (UID: "48c18077-569a-4bbd-ace2-64dbf06bcbee"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.191025 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-inventory" (OuterVolumeSpecName: "inventory") pod "48c18077-569a-4bbd-ace2-64dbf06bcbee" (UID: "48c18077-569a-4bbd-ace2-64dbf06bcbee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.201742 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "48c18077-569a-4bbd-ace2-64dbf06bcbee" (UID: "48c18077-569a-4bbd-ace2-64dbf06bcbee"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.201760 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "48c18077-569a-4bbd-ace2-64dbf06bcbee" (UID: "48c18077-569a-4bbd-ace2-64dbf06bcbee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.239559 4719 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.239618 4719 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.239633 4719 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.239644 4719 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.239656 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.239694 4719 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.239705 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.239719 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmm8q\" (UniqueName: \"kubernetes.io/projected/48c18077-569a-4bbd-ace2-64dbf06bcbee-kube-api-access-jmm8q\") on node \"crc\" DevicePath \"\"" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.239731 4719 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/48c18077-569a-4bbd-ace2-64dbf06bcbee-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.627641 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" event={"ID":"48c18077-569a-4bbd-ace2-64dbf06bcbee","Type":"ContainerDied","Data":"11484a88a77244482cff5caabe7639745a58c61bda25930618db3feca8db34f1"} Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.627686 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11484a88a77244482cff5caabe7639745a58c61bda25930618db3feca8db34f1" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.627739 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-hvqz9" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754067 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb"] Dec 15 13:04:56 crc kubenswrapper[4719]: E1215 13:04:56.754447 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerName="registry-server" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754463 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerName="registry-server" Dec 15 13:04:56 crc kubenswrapper[4719]: E1215 13:04:56.754482 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerName="extract-content" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754490 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerName="extract-content" Dec 15 13:04:56 crc kubenswrapper[4719]: E1215 13:04:56.754522 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerName="extract-utilities" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754531 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerName="extract-utilities" Dec 15 13:04:56 crc kubenswrapper[4719]: E1215 13:04:56.754547 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerName="registry-server" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754555 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerName="registry-server" Dec 15 13:04:56 crc kubenswrapper[4719]: E1215 13:04:56.754570 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerName="extract-content" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754576 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerName="extract-content" Dec 15 13:04:56 crc kubenswrapper[4719]: E1215 13:04:56.754591 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerName="extract-utilities" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754599 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerName="extract-utilities" Dec 15 13:04:56 crc kubenswrapper[4719]: E1215 13:04:56.754609 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c18077-569a-4bbd-ace2-64dbf06bcbee" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754615 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c18077-569a-4bbd-ace2-64dbf06bcbee" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754784 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="56dfcdfa-9677-498d-8f13-eaf8cabc48e3" containerName="registry-server" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754802 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee0096c1-a669-470d-ac7e-cab99afd927a" containerName="registry-server" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.754813 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c18077-569a-4bbd-ace2-64dbf06bcbee" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.755455 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.758589 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.758604 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.758846 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p854z" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.759086 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.759340 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.783002 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb"] Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.853161 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.853277 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.853448 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.853483 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.853565 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.853615 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.853650 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c2vr\" (UniqueName: \"kubernetes.io/projected/e70ef11b-7e5a-47c9-9f63-15dd69326073-kube-api-access-5c2vr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.955614 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.955657 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.955691 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c2vr\" (UniqueName: \"kubernetes.io/projected/e70ef11b-7e5a-47c9-9f63-15dd69326073-kube-api-access-5c2vr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.955723 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.955774 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.955849 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.955890 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.959517 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.959842 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.959907 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.960796 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.960874 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.961211 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:56 crc kubenswrapper[4719]: I1215 13:04:56.972677 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c2vr\" (UniqueName: \"kubernetes.io/projected/e70ef11b-7e5a-47c9-9f63-15dd69326073-kube-api-access-5c2vr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:57 crc kubenswrapper[4719]: I1215 13:04:57.081067 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:04:57 crc kubenswrapper[4719]: I1215 13:04:57.763080 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb"] Dec 15 13:04:58 crc kubenswrapper[4719]: I1215 13:04:58.646076 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" event={"ID":"e70ef11b-7e5a-47c9-9f63-15dd69326073","Type":"ContainerStarted","Data":"2c1776426070f49df69c0b1cf1f7cba5d8db8a50d9db3ae0a1e2489a634c61dd"} Dec 15 13:05:00 crc kubenswrapper[4719]: I1215 13:05:00.672838 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" event={"ID":"e70ef11b-7e5a-47c9-9f63-15dd69326073","Type":"ContainerStarted","Data":"6202113c9ff21908ad35c03138c55d8d66bebbe6934bf91d0a22cce4b0e66591"} Dec 15 13:05:00 crc kubenswrapper[4719]: I1215 13:05:00.699249 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" podStartSLOduration=2.954320445 podStartE2EDuration="4.699226305s" podCreationTimestamp="2025-12-15 13:04:56 +0000 UTC" firstStartedPulling="2025-12-15 13:04:57.770399869 +0000 UTC m=+2858.712692899" lastFinishedPulling="2025-12-15 13:04:59.515305729 +0000 UTC m=+2860.457598759" observedRunningTime="2025-12-15 13:05:00.689821401 +0000 UTC m=+2861.632114431" watchObservedRunningTime="2025-12-15 13:05:00.699226305 +0000 UTC m=+2861.641519345" Dec 15 13:06:21 crc kubenswrapper[4719]: I1215 13:06:21.357412 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:06:21 crc kubenswrapper[4719]: I1215 13:06:21.357994 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:06:34 crc kubenswrapper[4719]: I1215 13:06:34.857879 4719 scope.go:117] "RemoveContainer" containerID="308630120fe5812beeded8d7e7513fbea862871194e8e64e261cb245a3abbea2" Dec 15 13:06:34 crc kubenswrapper[4719]: I1215 13:06:34.890168 4719 scope.go:117] "RemoveContainer" containerID="9d6cdbffa95611910570c1a6448cf1924c18f18801e1a0d09f7f6ab126e90128" Dec 15 13:06:34 crc kubenswrapper[4719]: I1215 13:06:34.937216 4719 scope.go:117] "RemoveContainer" containerID="b0cf77812a3a9b6f8f4aafd49709c6eb09270fd14098e8d3752722708a8070bd" Dec 15 13:06:51 crc kubenswrapper[4719]: I1215 13:06:51.356481 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:06:51 crc kubenswrapper[4719]: I1215 13:06:51.357034 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:07:21 crc kubenswrapper[4719]: I1215 13:07:21.357220 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:07:21 crc kubenswrapper[4719]: I1215 13:07:21.358829 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:07:21 crc kubenswrapper[4719]: I1215 13:07:21.358963 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 13:07:21 crc kubenswrapper[4719]: I1215 13:07:21.359645 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 13:07:21 crc kubenswrapper[4719]: I1215 13:07:21.359994 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" gracePeriod=600 Dec 15 13:07:21 crc kubenswrapper[4719]: E1215 13:07:21.496852 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:07:22 crc kubenswrapper[4719]: I1215 13:07:22.005567 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" exitCode=0 Dec 15 13:07:22 crc kubenswrapper[4719]: I1215 13:07:22.005611 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d"} Dec 15 13:07:22 crc kubenswrapper[4719]: I1215 13:07:22.005643 4719 scope.go:117] "RemoveContainer" containerID="18ecba1ded68c775cb4acec3f16c1a4fc7cc2e424616603b49749c3ef16152b6" Dec 15 13:07:22 crc kubenswrapper[4719]: I1215 13:07:22.006549 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:07:22 crc kubenswrapper[4719]: E1215 13:07:22.007080 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:07:36 crc kubenswrapper[4719]: I1215 13:07:36.476558 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:07:36 crc kubenswrapper[4719]: E1215 13:07:36.477591 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:07:51 crc kubenswrapper[4719]: I1215 13:07:51.476310 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:07:51 crc kubenswrapper[4719]: E1215 13:07:51.476966 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:08:04 crc kubenswrapper[4719]: I1215 13:08:04.476346 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:08:04 crc kubenswrapper[4719]: E1215 13:08:04.477066 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:08:17 crc kubenswrapper[4719]: I1215 13:08:17.476483 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:08:17 crc kubenswrapper[4719]: E1215 13:08:17.477356 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:08:30 crc kubenswrapper[4719]: I1215 13:08:30.623037 4719 generic.go:334] "Generic (PLEG): container finished" podID="e70ef11b-7e5a-47c9-9f63-15dd69326073" containerID="6202113c9ff21908ad35c03138c55d8d66bebbe6934bf91d0a22cce4b0e66591" exitCode=0 Dec 15 13:08:30 crc kubenswrapper[4719]: I1215 13:08:30.623142 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" event={"ID":"e70ef11b-7e5a-47c9-9f63-15dd69326073","Type":"ContainerDied","Data":"6202113c9ff21908ad35c03138c55d8d66bebbe6934bf91d0a22cce4b0e66591"} Dec 15 13:08:31 crc kubenswrapper[4719]: I1215 13:08:31.478222 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:08:31 crc kubenswrapper[4719]: E1215 13:08:31.478812 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.004288 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.056564 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-0\") pod \"e70ef11b-7e5a-47c9-9f63-15dd69326073\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.056647 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-telemetry-combined-ca-bundle\") pod \"e70ef11b-7e5a-47c9-9f63-15dd69326073\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.056702 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-inventory\") pod \"e70ef11b-7e5a-47c9-9f63-15dd69326073\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.056760 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-2\") pod \"e70ef11b-7e5a-47c9-9f63-15dd69326073\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.056787 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-1\") pod \"e70ef11b-7e5a-47c9-9f63-15dd69326073\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.057252 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ssh-key\") pod \"e70ef11b-7e5a-47c9-9f63-15dd69326073\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.057746 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c2vr\" (UniqueName: \"kubernetes.io/projected/e70ef11b-7e5a-47c9-9f63-15dd69326073-kube-api-access-5c2vr\") pod \"e70ef11b-7e5a-47c9-9f63-15dd69326073\" (UID: \"e70ef11b-7e5a-47c9-9f63-15dd69326073\") " Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.063351 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "e70ef11b-7e5a-47c9-9f63-15dd69326073" (UID: "e70ef11b-7e5a-47c9-9f63-15dd69326073"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.083737 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e70ef11b-7e5a-47c9-9f63-15dd69326073-kube-api-access-5c2vr" (OuterVolumeSpecName: "kube-api-access-5c2vr") pod "e70ef11b-7e5a-47c9-9f63-15dd69326073" (UID: "e70ef11b-7e5a-47c9-9f63-15dd69326073"). InnerVolumeSpecName "kube-api-access-5c2vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.095714 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "e70ef11b-7e5a-47c9-9f63-15dd69326073" (UID: "e70ef11b-7e5a-47c9-9f63-15dd69326073"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.100147 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-inventory" (OuterVolumeSpecName: "inventory") pod "e70ef11b-7e5a-47c9-9f63-15dd69326073" (UID: "e70ef11b-7e5a-47c9-9f63-15dd69326073"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.101047 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "e70ef11b-7e5a-47c9-9f63-15dd69326073" (UID: "e70ef11b-7e5a-47c9-9f63-15dd69326073"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.115172 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e70ef11b-7e5a-47c9-9f63-15dd69326073" (UID: "e70ef11b-7e5a-47c9-9f63-15dd69326073"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.115454 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "e70ef11b-7e5a-47c9-9f63-15dd69326073" (UID: "e70ef11b-7e5a-47c9-9f63-15dd69326073"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.160702 4719 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-inventory\") on node \"crc\" DevicePath \"\"" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.160738 4719 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.160750 4719 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.160760 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.160770 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c2vr\" (UniqueName: \"kubernetes.io/projected/e70ef11b-7e5a-47c9-9f63-15dd69326073-kube-api-access-5c2vr\") on node \"crc\" DevicePath \"\"" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.160780 4719 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.160807 4719 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70ef11b-7e5a-47c9-9f63-15dd69326073-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.642048 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" event={"ID":"e70ef11b-7e5a-47c9-9f63-15dd69326073","Type":"ContainerDied","Data":"2c1776426070f49df69c0b1cf1f7cba5d8db8a50d9db3ae0a1e2489a634c61dd"} Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.642093 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c1776426070f49df69c0b1cf1f7cba5d8db8a50d9db3ae0a1e2489a634c61dd" Dec 15 13:08:32 crc kubenswrapper[4719]: I1215 13:08:32.642101 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb" Dec 15 13:08:42 crc kubenswrapper[4719]: I1215 13:08:42.478095 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:08:42 crc kubenswrapper[4719]: E1215 13:08:42.479243 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:08:56 crc kubenswrapper[4719]: I1215 13:08:56.476161 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:08:56 crc kubenswrapper[4719]: E1215 13:08:56.477090 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:09:11 crc kubenswrapper[4719]: I1215 13:09:11.476949 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:09:11 crc kubenswrapper[4719]: E1215 13:09:11.477536 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:09:23 crc kubenswrapper[4719]: I1215 13:09:23.476882 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:09:23 crc kubenswrapper[4719]: E1215 13:09:23.477558 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.056044 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 15 13:09:34 crc kubenswrapper[4719]: E1215 13:09:34.056919 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70ef11b-7e5a-47c9-9f63-15dd69326073" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.056933 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70ef11b-7e5a-47c9-9f63-15dd69326073" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.057101 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="e70ef11b-7e5a-47c9-9f63-15dd69326073" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.057673 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.059587 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.066208 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mlh57" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.066208 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.066558 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.066687 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.066834 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-config-data\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.066914 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.078800 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.168648 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.168695 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.168776 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.168800 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.168830 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.168897 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-config-data\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.168922 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.169003 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.169021 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvsjq\" (UniqueName: \"kubernetes.io/projected/8b81902c-c979-4e03-a40f-cac54a63d6db-kube-api-access-zvsjq\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.169799 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.170025 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-config-data\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.175316 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.270745 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.270787 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvsjq\" (UniqueName: \"kubernetes.io/projected/8b81902c-c979-4e03-a40f-cac54a63d6db-kube-api-access-zvsjq\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.270818 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.270834 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.270900 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.270929 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.271399 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.271436 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.272740 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.274579 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.280701 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.301118 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvsjq\" (UniqueName: \"kubernetes.io/projected/8b81902c-c979-4e03-a40f-cac54a63d6db-kube-api-access-zvsjq\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.306062 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.377764 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.478713 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:09:34 crc kubenswrapper[4719]: E1215 13:09:34.479294 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.841535 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 15 13:09:34 crc kubenswrapper[4719]: W1215 13:09:34.847079 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b81902c_c979_4e03_a40f_cac54a63d6db.slice/crio-ea5b6876af2f72484e40690c8757871aeaa90838994f480b0ee02c341864f907 WatchSource:0}: Error finding container ea5b6876af2f72484e40690c8757871aeaa90838994f480b0ee02c341864f907: Status 404 returned error can't find the container with id ea5b6876af2f72484e40690c8757871aeaa90838994f480b0ee02c341864f907 Dec 15 13:09:34 crc kubenswrapper[4719]: I1215 13:09:34.849533 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 13:09:35 crc kubenswrapper[4719]: I1215 13:09:35.169976 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8b81902c-c979-4e03-a40f-cac54a63d6db","Type":"ContainerStarted","Data":"ea5b6876af2f72484e40690c8757871aeaa90838994f480b0ee02c341864f907"} Dec 15 13:09:47 crc kubenswrapper[4719]: I1215 13:09:47.477685 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:09:47 crc kubenswrapper[4719]: E1215 13:09:47.478592 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:09:59 crc kubenswrapper[4719]: I1215 13:09:59.476489 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:09:59 crc kubenswrapper[4719]: E1215 13:09:59.477290 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:10:10 crc kubenswrapper[4719]: E1215 13:10:10.264252 4719 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 15 13:10:10 crc kubenswrapper[4719]: E1215 13:10:10.264837 4719 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zvsjq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(8b81902c-c979-4e03-a40f-cac54a63d6db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 15 13:10:10 crc kubenswrapper[4719]: E1215 13:10:10.266095 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="8b81902c-c979-4e03-a40f-cac54a63d6db" Dec 15 13:10:10 crc kubenswrapper[4719]: E1215 13:10:10.488929 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="8b81902c-c979-4e03-a40f-cac54a63d6db" Dec 15 13:10:11 crc kubenswrapper[4719]: I1215 13:10:11.476716 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:10:11 crc kubenswrapper[4719]: E1215 13:10:11.477369 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:10:22 crc kubenswrapper[4719]: I1215 13:10:22.476424 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:10:22 crc kubenswrapper[4719]: E1215 13:10:22.477265 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:10:23 crc kubenswrapper[4719]: I1215 13:10:23.978185 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 15 13:10:25 crc kubenswrapper[4719]: I1215 13:10:25.617160 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8b81902c-c979-4e03-a40f-cac54a63d6db","Type":"ContainerStarted","Data":"5ad12e1d824e57bb6f5d03028c8ed11cc6b0b0bebce7e6a040bceb64737fdda2"} Dec 15 13:10:25 crc kubenswrapper[4719]: I1215 13:10:25.643506 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.5172287730000003 podStartE2EDuration="52.643491594s" podCreationTimestamp="2025-12-15 13:09:33 +0000 UTC" firstStartedPulling="2025-12-15 13:09:34.849221142 +0000 UTC m=+3135.791514172" lastFinishedPulling="2025-12-15 13:10:23.975483963 +0000 UTC m=+3184.917776993" observedRunningTime="2025-12-15 13:10:25.64143565 +0000 UTC m=+3186.583728690" watchObservedRunningTime="2025-12-15 13:10:25.643491594 +0000 UTC m=+3186.585784624" Dec 15 13:10:33 crc kubenswrapper[4719]: I1215 13:10:33.476306 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:10:33 crc kubenswrapper[4719]: E1215 13:10:33.477079 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:10:45 crc kubenswrapper[4719]: I1215 13:10:45.477807 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:10:45 crc kubenswrapper[4719]: E1215 13:10:45.478786 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:11:00 crc kubenswrapper[4719]: I1215 13:11:00.476295 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:11:00 crc kubenswrapper[4719]: E1215 13:11:00.477159 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:11:13 crc kubenswrapper[4719]: I1215 13:11:13.476322 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:11:13 crc kubenswrapper[4719]: E1215 13:11:13.477167 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:11:24 crc kubenswrapper[4719]: I1215 13:11:24.475840 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:11:24 crc kubenswrapper[4719]: E1215 13:11:24.477043 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:11:37 crc kubenswrapper[4719]: I1215 13:11:37.476040 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:11:37 crc kubenswrapper[4719]: E1215 13:11:37.476616 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:11:44 crc kubenswrapper[4719]: I1215 13:11:44.985534 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rxv5c"] Dec 15 13:11:44 crc kubenswrapper[4719]: I1215 13:11:44.988307 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.046152 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxv5c"] Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.162084 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-catalog-content\") pod \"redhat-marketplace-rxv5c\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.162135 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-utilities\") pod \"redhat-marketplace-rxv5c\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.162206 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v62mx\" (UniqueName: \"kubernetes.io/projected/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-kube-api-access-v62mx\") pod \"redhat-marketplace-rxv5c\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.176237 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6qxhz"] Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.178056 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.189783 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6qxhz"] Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.263488 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v62mx\" (UniqueName: \"kubernetes.io/projected/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-kube-api-access-v62mx\") pod \"redhat-marketplace-rxv5c\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.263652 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-catalog-content\") pod \"redhat-marketplace-rxv5c\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.263674 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-utilities\") pod \"redhat-marketplace-rxv5c\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.264210 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-utilities\") pod \"redhat-marketplace-rxv5c\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.264290 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-catalog-content\") pod \"redhat-marketplace-rxv5c\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.290472 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v62mx\" (UniqueName: \"kubernetes.io/projected/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-kube-api-access-v62mx\") pod \"redhat-marketplace-rxv5c\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.308967 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.373592 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-catalog-content\") pod \"community-operators-6qxhz\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.374059 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-utilities\") pod \"community-operators-6qxhz\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.374114 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p67m\" (UniqueName: \"kubernetes.io/projected/351aab48-2379-4481-8752-82bd22525194-kube-api-access-2p67m\") pod \"community-operators-6qxhz\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.480764 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-utilities\") pod \"community-operators-6qxhz\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.480817 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p67m\" (UniqueName: \"kubernetes.io/projected/351aab48-2379-4481-8752-82bd22525194-kube-api-access-2p67m\") pod \"community-operators-6qxhz\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.480893 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-catalog-content\") pod \"community-operators-6qxhz\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.481409 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-catalog-content\") pod \"community-operators-6qxhz\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.481709 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-utilities\") pod \"community-operators-6qxhz\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.519967 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p67m\" (UniqueName: \"kubernetes.io/projected/351aab48-2379-4481-8752-82bd22525194-kube-api-access-2p67m\") pod \"community-operators-6qxhz\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:45 crc kubenswrapper[4719]: I1215 13:11:45.795966 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:11:46 crc kubenswrapper[4719]: I1215 13:11:46.214500 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxv5c"] Dec 15 13:11:46 crc kubenswrapper[4719]: I1215 13:11:46.359074 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxv5c" event={"ID":"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb","Type":"ContainerStarted","Data":"900ae911b83d8d32007f0bbcc03bf72e7adf7b6afff083ad7628daa80f0c0850"} Dec 15 13:11:46 crc kubenswrapper[4719]: I1215 13:11:46.605669 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6qxhz"] Dec 15 13:11:47 crc kubenswrapper[4719]: I1215 13:11:47.376262 4719 generic.go:334] "Generic (PLEG): container finished" podID="351aab48-2379-4481-8752-82bd22525194" containerID="de90b6a654988b0ebfc58a7e68010fc6c49bb4a275372b999a11f7302682e92f" exitCode=0 Dec 15 13:11:47 crc kubenswrapper[4719]: I1215 13:11:47.376459 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qxhz" event={"ID":"351aab48-2379-4481-8752-82bd22525194","Type":"ContainerDied","Data":"de90b6a654988b0ebfc58a7e68010fc6c49bb4a275372b999a11f7302682e92f"} Dec 15 13:11:47 crc kubenswrapper[4719]: I1215 13:11:47.376558 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qxhz" event={"ID":"351aab48-2379-4481-8752-82bd22525194","Type":"ContainerStarted","Data":"9a726c7aa0c3e940ea38ed5189087360af30c0d8fff0f82402f938d275fe1896"} Dec 15 13:11:47 crc kubenswrapper[4719]: I1215 13:11:47.379083 4719 generic.go:334] "Generic (PLEG): container finished" podID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerID="86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce" exitCode=0 Dec 15 13:11:47 crc kubenswrapper[4719]: I1215 13:11:47.379113 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxv5c" event={"ID":"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb","Type":"ContainerDied","Data":"86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce"} Dec 15 13:11:47 crc kubenswrapper[4719]: I1215 13:11:47.979632 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4fln2"] Dec 15 13:11:47 crc kubenswrapper[4719]: I1215 13:11:47.997618 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.023105 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4fln2"] Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.139586 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-catalog-content\") pod \"redhat-operators-4fln2\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.139982 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5chqr\" (UniqueName: \"kubernetes.io/projected/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-kube-api-access-5chqr\") pod \"redhat-operators-4fln2\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.140111 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-utilities\") pod \"redhat-operators-4fln2\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.241804 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5chqr\" (UniqueName: \"kubernetes.io/projected/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-kube-api-access-5chqr\") pod \"redhat-operators-4fln2\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.241890 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-utilities\") pod \"redhat-operators-4fln2\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.241934 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-catalog-content\") pod \"redhat-operators-4fln2\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.242573 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-catalog-content\") pod \"redhat-operators-4fln2\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.242704 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-utilities\") pod \"redhat-operators-4fln2\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.261077 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5chqr\" (UniqueName: \"kubernetes.io/projected/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-kube-api-access-5chqr\") pod \"redhat-operators-4fln2\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.340188 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:11:48 crc kubenswrapper[4719]: I1215 13:11:48.881724 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4fln2"] Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.413638 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxv5c" event={"ID":"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb","Type":"ContainerStarted","Data":"f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77"} Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.415207 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fln2" event={"ID":"cab1f22d-ecb1-470e-8b10-61b4b0f8817d","Type":"ContainerStarted","Data":"b0da2918ecd2f4d28274454225403992bdf39452614309c824b9778db3d78a08"} Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.417422 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qxhz" event={"ID":"351aab48-2379-4481-8752-82bd22525194","Type":"ContainerStarted","Data":"367c5265ba3232169b495df6d41f8a24c33ac2c91fd9ecc0a03a11e308ebd1d8"} Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.469019 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2nknk"] Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.471486 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.541545 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2nknk"] Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.571440 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scq5t\" (UniqueName: \"kubernetes.io/projected/ccefa2e7-7772-4804-977b-f6600bbfe02d-kube-api-access-scq5t\") pod \"certified-operators-2nknk\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.571671 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-catalog-content\") pod \"certified-operators-2nknk\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.571700 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-utilities\") pod \"certified-operators-2nknk\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.673892 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-catalog-content\") pod \"certified-operators-2nknk\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.673954 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-utilities\") pod \"certified-operators-2nknk\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.674007 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scq5t\" (UniqueName: \"kubernetes.io/projected/ccefa2e7-7772-4804-977b-f6600bbfe02d-kube-api-access-scq5t\") pod \"certified-operators-2nknk\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.674407 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-catalog-content\") pod \"certified-operators-2nknk\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.674798 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-utilities\") pod \"certified-operators-2nknk\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.695456 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scq5t\" (UniqueName: \"kubernetes.io/projected/ccefa2e7-7772-4804-977b-f6600bbfe02d-kube-api-access-scq5t\") pod \"certified-operators-2nknk\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:49 crc kubenswrapper[4719]: I1215 13:11:49.865333 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:11:50 crc kubenswrapper[4719]: I1215 13:11:50.428007 4719 generic.go:334] "Generic (PLEG): container finished" podID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerID="f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77" exitCode=0 Dec 15 13:11:50 crc kubenswrapper[4719]: I1215 13:11:50.428106 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxv5c" event={"ID":"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb","Type":"ContainerDied","Data":"f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77"} Dec 15 13:11:50 crc kubenswrapper[4719]: I1215 13:11:50.576416 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2nknk"] Dec 15 13:11:51 crc kubenswrapper[4719]: I1215 13:11:51.437992 4719 generic.go:334] "Generic (PLEG): container finished" podID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerID="06ece604772c2accaf07157da0c3ab7916b3e1c7d2ca20fbc7e77809e725cc61" exitCode=0 Dec 15 13:11:51 crc kubenswrapper[4719]: I1215 13:11:51.438271 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fln2" event={"ID":"cab1f22d-ecb1-470e-8b10-61b4b0f8817d","Type":"ContainerDied","Data":"06ece604772c2accaf07157da0c3ab7916b3e1c7d2ca20fbc7e77809e725cc61"} Dec 15 13:11:51 crc kubenswrapper[4719]: I1215 13:11:51.441273 4719 generic.go:334] "Generic (PLEG): container finished" podID="351aab48-2379-4481-8752-82bd22525194" containerID="367c5265ba3232169b495df6d41f8a24c33ac2c91fd9ecc0a03a11e308ebd1d8" exitCode=0 Dec 15 13:11:51 crc kubenswrapper[4719]: I1215 13:11:51.441343 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qxhz" event={"ID":"351aab48-2379-4481-8752-82bd22525194","Type":"ContainerDied","Data":"367c5265ba3232169b495df6d41f8a24c33ac2c91fd9ecc0a03a11e308ebd1d8"} Dec 15 13:11:51 crc kubenswrapper[4719]: I1215 13:11:51.442897 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2nknk" event={"ID":"ccefa2e7-7772-4804-977b-f6600bbfe02d","Type":"ContainerStarted","Data":"1fdd2bc2b171eb32801a88ebdb88ee91c027f151ee43ca85f8cca844eb2dcd8e"} Dec 15 13:11:52 crc kubenswrapper[4719]: I1215 13:11:52.453129 4719 generic.go:334] "Generic (PLEG): container finished" podID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerID="41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f" exitCode=0 Dec 15 13:11:52 crc kubenswrapper[4719]: I1215 13:11:52.453201 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2nknk" event={"ID":"ccefa2e7-7772-4804-977b-f6600bbfe02d","Type":"ContainerDied","Data":"41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f"} Dec 15 13:11:52 crc kubenswrapper[4719]: I1215 13:11:52.475884 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:11:52 crc kubenswrapper[4719]: E1215 13:11:52.476142 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:11:59 crc kubenswrapper[4719]: I1215 13:11:59.515906 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxv5c" event={"ID":"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb","Type":"ContainerStarted","Data":"f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b"} Dec 15 13:11:59 crc kubenswrapper[4719]: I1215 13:11:59.545605 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rxv5c" podStartSLOduration=5.755527956 podStartE2EDuration="15.545589381s" podCreationTimestamp="2025-12-15 13:11:44 +0000 UTC" firstStartedPulling="2025-12-15 13:11:47.382200451 +0000 UTC m=+3268.324493501" lastFinishedPulling="2025-12-15 13:11:57.172261896 +0000 UTC m=+3278.114554926" observedRunningTime="2025-12-15 13:11:59.538589742 +0000 UTC m=+3280.480882782" watchObservedRunningTime="2025-12-15 13:11:59.545589381 +0000 UTC m=+3280.487882411" Dec 15 13:12:00 crc kubenswrapper[4719]: I1215 13:12:00.527303 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qxhz" event={"ID":"351aab48-2379-4481-8752-82bd22525194","Type":"ContainerStarted","Data":"0935abf0d605b70b6c3f0e7c3fd3987830e6d156197edbb24b06e570e852c368"} Dec 15 13:12:00 crc kubenswrapper[4719]: I1215 13:12:00.552330 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6qxhz" podStartSLOduration=3.034856155 podStartE2EDuration="15.552311647s" podCreationTimestamp="2025-12-15 13:11:45 +0000 UTC" firstStartedPulling="2025-12-15 13:11:47.377630398 +0000 UTC m=+3268.319923428" lastFinishedPulling="2025-12-15 13:11:59.89508589 +0000 UTC m=+3280.837378920" observedRunningTime="2025-12-15 13:12:00.551279174 +0000 UTC m=+3281.493572204" watchObservedRunningTime="2025-12-15 13:12:00.552311647 +0000 UTC m=+3281.494604687" Dec 15 13:12:02 crc kubenswrapper[4719]: I1215 13:12:02.545483 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fln2" event={"ID":"cab1f22d-ecb1-470e-8b10-61b4b0f8817d","Type":"ContainerStarted","Data":"814739b420072df825f9394388cfccd76fa74cfe6e483208a287e6ab49f535b6"} Dec 15 13:12:02 crc kubenswrapper[4719]: I1215 13:12:02.549216 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2nknk" event={"ID":"ccefa2e7-7772-4804-977b-f6600bbfe02d","Type":"ContainerStarted","Data":"59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1"} Dec 15 13:12:05 crc kubenswrapper[4719]: I1215 13:12:05.310447 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:12:05 crc kubenswrapper[4719]: I1215 13:12:05.310845 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:12:05 crc kubenswrapper[4719]: I1215 13:12:05.797297 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:12:05 crc kubenswrapper[4719]: I1215 13:12:05.797622 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:12:06 crc kubenswrapper[4719]: I1215 13:12:06.366439 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-rxv5c" podUID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerName="registry-server" probeResult="failure" output=< Dec 15 13:12:06 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 13:12:06 crc kubenswrapper[4719]: > Dec 15 13:12:06 crc kubenswrapper[4719]: I1215 13:12:06.591601 4719 generic.go:334] "Generic (PLEG): container finished" podID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerID="59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1" exitCode=0 Dec 15 13:12:06 crc kubenswrapper[4719]: I1215 13:12:06.591640 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2nknk" event={"ID":"ccefa2e7-7772-4804-977b-f6600bbfe02d","Type":"ContainerDied","Data":"59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1"} Dec 15 13:12:06 crc kubenswrapper[4719]: I1215 13:12:06.838811 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-6qxhz" podUID="351aab48-2379-4481-8752-82bd22525194" containerName="registry-server" probeResult="failure" output=< Dec 15 13:12:06 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 13:12:06 crc kubenswrapper[4719]: > Dec 15 13:12:07 crc kubenswrapper[4719]: I1215 13:12:07.476737 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:12:07 crc kubenswrapper[4719]: E1215 13:12:07.477219 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:12:08 crc kubenswrapper[4719]: I1215 13:12:08.621158 4719 generic.go:334] "Generic (PLEG): container finished" podID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerID="814739b420072df825f9394388cfccd76fa74cfe6e483208a287e6ab49f535b6" exitCode=0 Dec 15 13:12:08 crc kubenswrapper[4719]: I1215 13:12:08.621591 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fln2" event={"ID":"cab1f22d-ecb1-470e-8b10-61b4b0f8817d","Type":"ContainerDied","Data":"814739b420072df825f9394388cfccd76fa74cfe6e483208a287e6ab49f535b6"} Dec 15 13:12:08 crc kubenswrapper[4719]: I1215 13:12:08.631877 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2nknk" event={"ID":"ccefa2e7-7772-4804-977b-f6600bbfe02d","Type":"ContainerStarted","Data":"ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d"} Dec 15 13:12:08 crc kubenswrapper[4719]: I1215 13:12:08.689117 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2nknk" podStartSLOduration=6.362048596 podStartE2EDuration="20.689097303s" podCreationTimestamp="2025-12-15 13:11:48 +0000 UTC" firstStartedPulling="2025-12-15 13:11:53.464518121 +0000 UTC m=+3274.406811151" lastFinishedPulling="2025-12-15 13:12:07.791566828 +0000 UTC m=+3288.733859858" observedRunningTime="2025-12-15 13:12:08.670683126 +0000 UTC m=+3289.612976156" watchObservedRunningTime="2025-12-15 13:12:08.689097303 +0000 UTC m=+3289.631390333" Dec 15 13:12:09 crc kubenswrapper[4719]: I1215 13:12:09.644206 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fln2" event={"ID":"cab1f22d-ecb1-470e-8b10-61b4b0f8817d","Type":"ContainerStarted","Data":"2500dd95f1a85a7b105e844153217ad91976d3367c4b718f4abf9ff3187ec7ee"} Dec 15 13:12:09 crc kubenswrapper[4719]: I1215 13:12:09.666415 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4fln2" podStartSLOduration=5.740466289 podStartE2EDuration="22.666393507s" podCreationTimestamp="2025-12-15 13:11:47 +0000 UTC" firstStartedPulling="2025-12-15 13:11:52.455227534 +0000 UTC m=+3273.397520564" lastFinishedPulling="2025-12-15 13:12:09.381154752 +0000 UTC m=+3290.323447782" observedRunningTime="2025-12-15 13:12:09.660982417 +0000 UTC m=+3290.603275447" watchObservedRunningTime="2025-12-15 13:12:09.666393507 +0000 UTC m=+3290.608686537" Dec 15 13:12:09 crc kubenswrapper[4719]: I1215 13:12:09.866185 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:12:09 crc kubenswrapper[4719]: I1215 13:12:09.866359 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:12:10 crc kubenswrapper[4719]: I1215 13:12:10.915996 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-2nknk" podUID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerName="registry-server" probeResult="failure" output=< Dec 15 13:12:10 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 13:12:10 crc kubenswrapper[4719]: > Dec 15 13:12:15 crc kubenswrapper[4719]: I1215 13:12:15.360665 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:12:15 crc kubenswrapper[4719]: I1215 13:12:15.414921 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:12:15 crc kubenswrapper[4719]: I1215 13:12:15.856121 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:12:15 crc kubenswrapper[4719]: I1215 13:12:15.921764 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:12:16 crc kubenswrapper[4719]: I1215 13:12:16.388359 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxv5c"] Dec 15 13:12:16 crc kubenswrapper[4719]: I1215 13:12:16.699566 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rxv5c" podUID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerName="registry-server" containerID="cri-o://f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b" gracePeriod=2 Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.512164 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.634020 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-catalog-content\") pod \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.634075 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v62mx\" (UniqueName: \"kubernetes.io/projected/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-kube-api-access-v62mx\") pod \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.635204 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-utilities\") pod \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\" (UID: \"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb\") " Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.635496 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-utilities" (OuterVolumeSpecName: "utilities") pod "6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" (UID: "6e283d28-1b64-4b7f-aac2-b3ad186c6fcb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.635699 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.643234 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-kube-api-access-v62mx" (OuterVolumeSpecName: "kube-api-access-v62mx") pod "6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" (UID: "6e283d28-1b64-4b7f-aac2-b3ad186c6fcb"). InnerVolumeSpecName "kube-api-access-v62mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.656849 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" (UID: "6e283d28-1b64-4b7f-aac2-b3ad186c6fcb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.723228 4719 generic.go:334] "Generic (PLEG): container finished" podID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerID="f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b" exitCode=0 Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.723287 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxv5c" event={"ID":"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb","Type":"ContainerDied","Data":"f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b"} Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.723318 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rxv5c" event={"ID":"6e283d28-1b64-4b7f-aac2-b3ad186c6fcb","Type":"ContainerDied","Data":"900ae911b83d8d32007f0bbcc03bf72e7adf7b6afff083ad7628daa80f0c0850"} Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.723350 4719 scope.go:117] "RemoveContainer" containerID="f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.723339 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rxv5c" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.737171 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.737404 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v62mx\" (UniqueName: \"kubernetes.io/projected/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb-kube-api-access-v62mx\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.752086 4719 scope.go:117] "RemoveContainer" containerID="f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.775907 4719 scope.go:117] "RemoveContainer" containerID="86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.795189 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxv5c"] Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.803926 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rxv5c"] Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.823809 4719 scope.go:117] "RemoveContainer" containerID="f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b" Dec 15 13:12:17 crc kubenswrapper[4719]: E1215 13:12:17.824474 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b\": container with ID starting with f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b not found: ID does not exist" containerID="f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.824567 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b"} err="failed to get container status \"f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b\": rpc error: code = NotFound desc = could not find container \"f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b\": container with ID starting with f14b81bf11ddba683edc271084bb671e94364c3bd6699476ec28855b516acb3b not found: ID does not exist" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.824644 4719 scope.go:117] "RemoveContainer" containerID="f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77" Dec 15 13:12:17 crc kubenswrapper[4719]: E1215 13:12:17.824984 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77\": container with ID starting with f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77 not found: ID does not exist" containerID="f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.825082 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77"} err="failed to get container status \"f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77\": rpc error: code = NotFound desc = could not find container \"f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77\": container with ID starting with f5127acc5fff8cf86c02b782957fb32dd97a8c0a9d18877453b2b4fa7a6c8d77 not found: ID does not exist" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.825154 4719 scope.go:117] "RemoveContainer" containerID="86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce" Dec 15 13:12:17 crc kubenswrapper[4719]: E1215 13:12:17.826292 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce\": container with ID starting with 86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce not found: ID does not exist" containerID="86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce" Dec 15 13:12:17 crc kubenswrapper[4719]: I1215 13:12:17.826403 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce"} err="failed to get container status \"86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce\": rpc error: code = NotFound desc = could not find container \"86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce\": container with ID starting with 86935846bfc61d2f9927982b59124206d413f2ae9746c67da73c52edaee23dce not found: ID does not exist" Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.217130 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6qxhz"] Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.217474 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6qxhz" podUID="351aab48-2379-4481-8752-82bd22525194" containerName="registry-server" containerID="cri-o://0935abf0d605b70b6c3f0e7c3fd3987830e6d156197edbb24b06e570e852c368" gracePeriod=2 Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.340715 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.340748 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.734572 4719 generic.go:334] "Generic (PLEG): container finished" podID="351aab48-2379-4481-8752-82bd22525194" containerID="0935abf0d605b70b6c3f0e7c3fd3987830e6d156197edbb24b06e570e852c368" exitCode=0 Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.734650 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qxhz" event={"ID":"351aab48-2379-4481-8752-82bd22525194","Type":"ContainerDied","Data":"0935abf0d605b70b6c3f0e7c3fd3987830e6d156197edbb24b06e570e852c368"} Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.851514 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.965577 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-catalog-content\") pod \"351aab48-2379-4481-8752-82bd22525194\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.965791 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-utilities\") pod \"351aab48-2379-4481-8752-82bd22525194\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.965881 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p67m\" (UniqueName: \"kubernetes.io/projected/351aab48-2379-4481-8752-82bd22525194-kube-api-access-2p67m\") pod \"351aab48-2379-4481-8752-82bd22525194\" (UID: \"351aab48-2379-4481-8752-82bd22525194\") " Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.968131 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-utilities" (OuterVolumeSpecName: "utilities") pod "351aab48-2379-4481-8752-82bd22525194" (UID: "351aab48-2379-4481-8752-82bd22525194"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:12:18 crc kubenswrapper[4719]: I1215 13:12:18.972377 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/351aab48-2379-4481-8752-82bd22525194-kube-api-access-2p67m" (OuterVolumeSpecName: "kube-api-access-2p67m") pod "351aab48-2379-4481-8752-82bd22525194" (UID: "351aab48-2379-4481-8752-82bd22525194"). InnerVolumeSpecName "kube-api-access-2p67m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.041570 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "351aab48-2379-4481-8752-82bd22525194" (UID: "351aab48-2379-4481-8752-82bd22525194"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.068368 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p67m\" (UniqueName: \"kubernetes.io/projected/351aab48-2379-4481-8752-82bd22525194-kube-api-access-2p67m\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.068637 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.068651 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/351aab48-2379-4481-8752-82bd22525194-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.392316 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4fln2" podUID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerName="registry-server" probeResult="failure" output=< Dec 15 13:12:19 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 13:12:19 crc kubenswrapper[4719]: > Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.504778 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" path="/var/lib/kubelet/pods/6e283d28-1b64-4b7f-aac2-b3ad186c6fcb/volumes" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.746992 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qxhz" event={"ID":"351aab48-2379-4481-8752-82bd22525194","Type":"ContainerDied","Data":"9a726c7aa0c3e940ea38ed5189087360af30c0d8fff0f82402f938d275fe1896"} Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.747058 4719 scope.go:117] "RemoveContainer" containerID="0935abf0d605b70b6c3f0e7c3fd3987830e6d156197edbb24b06e570e852c368" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.747078 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qxhz" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.770499 4719 scope.go:117] "RemoveContainer" containerID="367c5265ba3232169b495df6d41f8a24c33ac2c91fd9ecc0a03a11e308ebd1d8" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.779029 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6qxhz"] Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.792359 4719 scope.go:117] "RemoveContainer" containerID="de90b6a654988b0ebfc58a7e68010fc6c49bb4a275372b999a11f7302682e92f" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.815004 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6qxhz"] Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.921454 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:12:19 crc kubenswrapper[4719]: I1215 13:12:19.981130 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:12:21 crc kubenswrapper[4719]: I1215 13:12:21.492709 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="351aab48-2379-4481-8752-82bd22525194" path="/var/lib/kubelet/pods/351aab48-2379-4481-8752-82bd22525194/volumes" Dec 15 13:12:22 crc kubenswrapper[4719]: I1215 13:12:22.476039 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:12:22 crc kubenswrapper[4719]: I1215 13:12:22.774444 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"3af3a7a5155192d69dc971d21bcb5ef1e92363efaa38759a082b973f30a65031"} Dec 15 13:12:22 crc kubenswrapper[4719]: I1215 13:12:22.787479 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2nknk"] Dec 15 13:12:22 crc kubenswrapper[4719]: I1215 13:12:22.787704 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2nknk" podUID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerName="registry-server" containerID="cri-o://ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d" gracePeriod=2 Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.469979 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.573183 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scq5t\" (UniqueName: \"kubernetes.io/projected/ccefa2e7-7772-4804-977b-f6600bbfe02d-kube-api-access-scq5t\") pod \"ccefa2e7-7772-4804-977b-f6600bbfe02d\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.573536 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-catalog-content\") pod \"ccefa2e7-7772-4804-977b-f6600bbfe02d\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.573679 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-utilities\") pod \"ccefa2e7-7772-4804-977b-f6600bbfe02d\" (UID: \"ccefa2e7-7772-4804-977b-f6600bbfe02d\") " Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.574105 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-utilities" (OuterVolumeSpecName: "utilities") pod "ccefa2e7-7772-4804-977b-f6600bbfe02d" (UID: "ccefa2e7-7772-4804-977b-f6600bbfe02d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.574556 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.578542 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccefa2e7-7772-4804-977b-f6600bbfe02d-kube-api-access-scq5t" (OuterVolumeSpecName: "kube-api-access-scq5t") pod "ccefa2e7-7772-4804-977b-f6600bbfe02d" (UID: "ccefa2e7-7772-4804-977b-f6600bbfe02d"). InnerVolumeSpecName "kube-api-access-scq5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.627743 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccefa2e7-7772-4804-977b-f6600bbfe02d" (UID: "ccefa2e7-7772-4804-977b-f6600bbfe02d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.677115 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scq5t\" (UniqueName: \"kubernetes.io/projected/ccefa2e7-7772-4804-977b-f6600bbfe02d-kube-api-access-scq5t\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.677155 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccefa2e7-7772-4804-977b-f6600bbfe02d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.788005 4719 generic.go:334] "Generic (PLEG): container finished" podID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerID="ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d" exitCode=0 Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.788044 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2nknk" event={"ID":"ccefa2e7-7772-4804-977b-f6600bbfe02d","Type":"ContainerDied","Data":"ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d"} Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.788072 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2nknk" event={"ID":"ccefa2e7-7772-4804-977b-f6600bbfe02d","Type":"ContainerDied","Data":"1fdd2bc2b171eb32801a88ebdb88ee91c027f151ee43ca85f8cca844eb2dcd8e"} Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.788090 4719 scope.go:117] "RemoveContainer" containerID="ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.788090 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2nknk" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.830639 4719 scope.go:117] "RemoveContainer" containerID="59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.858784 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2nknk"] Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.871631 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2nknk"] Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.873775 4719 scope.go:117] "RemoveContainer" containerID="41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.904664 4719 scope.go:117] "RemoveContainer" containerID="ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d" Dec 15 13:12:23 crc kubenswrapper[4719]: E1215 13:12:23.905325 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d\": container with ID starting with ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d not found: ID does not exist" containerID="ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.905384 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d"} err="failed to get container status \"ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d\": rpc error: code = NotFound desc = could not find container \"ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d\": container with ID starting with ee69f3fcd65785183c8cdcfb2b9d90ded46ccbcdb0ff841a24d722b7559de30d not found: ID does not exist" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.905420 4719 scope.go:117] "RemoveContainer" containerID="59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1" Dec 15 13:12:23 crc kubenswrapper[4719]: E1215 13:12:23.905993 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1\": container with ID starting with 59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1 not found: ID does not exist" containerID="59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.906036 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1"} err="failed to get container status \"59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1\": rpc error: code = NotFound desc = could not find container \"59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1\": container with ID starting with 59a7babe04c3787d81f3d50fc9044fb45d5ef70c55e66bc9114c08c2402aaaa1 not found: ID does not exist" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.906058 4719 scope.go:117] "RemoveContainer" containerID="41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f" Dec 15 13:12:23 crc kubenswrapper[4719]: E1215 13:12:23.906599 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f\": container with ID starting with 41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f not found: ID does not exist" containerID="41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f" Dec 15 13:12:23 crc kubenswrapper[4719]: I1215 13:12:23.906626 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f"} err="failed to get container status \"41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f\": rpc error: code = NotFound desc = could not find container \"41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f\": container with ID starting with 41e144caf241fa2d575bfef52ff20ef7a7c22301c530f51d4715ed659648116f not found: ID does not exist" Dec 15 13:12:25 crc kubenswrapper[4719]: I1215 13:12:25.490509 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccefa2e7-7772-4804-977b-f6600bbfe02d" path="/var/lib/kubelet/pods/ccefa2e7-7772-4804-977b-f6600bbfe02d/volumes" Dec 15 13:12:28 crc kubenswrapper[4719]: I1215 13:12:28.390371 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:12:28 crc kubenswrapper[4719]: I1215 13:12:28.445408 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:12:28 crc kubenswrapper[4719]: I1215 13:12:28.785397 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4fln2"] Dec 15 13:12:29 crc kubenswrapper[4719]: I1215 13:12:29.831498 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4fln2" podUID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerName="registry-server" containerID="cri-o://2500dd95f1a85a7b105e844153217ad91976d3367c4b718f4abf9ff3187ec7ee" gracePeriod=2 Dec 15 13:12:30 crc kubenswrapper[4719]: I1215 13:12:30.842628 4719 generic.go:334] "Generic (PLEG): container finished" podID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerID="2500dd95f1a85a7b105e844153217ad91976d3367c4b718f4abf9ff3187ec7ee" exitCode=0 Dec 15 13:12:30 crc kubenswrapper[4719]: I1215 13:12:30.842986 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fln2" event={"ID":"cab1f22d-ecb1-470e-8b10-61b4b0f8817d","Type":"ContainerDied","Data":"2500dd95f1a85a7b105e844153217ad91976d3367c4b718f4abf9ff3187ec7ee"} Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.019685 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.155400 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-catalog-content\") pod \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.156129 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5chqr\" (UniqueName: \"kubernetes.io/projected/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-kube-api-access-5chqr\") pod \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.156181 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-utilities\") pod \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\" (UID: \"cab1f22d-ecb1-470e-8b10-61b4b0f8817d\") " Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.156980 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-utilities" (OuterVolumeSpecName: "utilities") pod "cab1f22d-ecb1-470e-8b10-61b4b0f8817d" (UID: "cab1f22d-ecb1-470e-8b10-61b4b0f8817d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.157967 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.169537 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-kube-api-access-5chqr" (OuterVolumeSpecName: "kube-api-access-5chqr") pod "cab1f22d-ecb1-470e-8b10-61b4b0f8817d" (UID: "cab1f22d-ecb1-470e-8b10-61b4b0f8817d"). InnerVolumeSpecName "kube-api-access-5chqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.260182 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5chqr\" (UniqueName: \"kubernetes.io/projected/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-kube-api-access-5chqr\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.274295 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cab1f22d-ecb1-470e-8b10-61b4b0f8817d" (UID: "cab1f22d-ecb1-470e-8b10-61b4b0f8817d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.362247 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cab1f22d-ecb1-470e-8b10-61b4b0f8817d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.853634 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4fln2" event={"ID":"cab1f22d-ecb1-470e-8b10-61b4b0f8817d","Type":"ContainerDied","Data":"b0da2918ecd2f4d28274454225403992bdf39452614309c824b9778db3d78a08"} Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.853694 4719 scope.go:117] "RemoveContainer" containerID="2500dd95f1a85a7b105e844153217ad91976d3367c4b718f4abf9ff3187ec7ee" Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.853724 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4fln2" Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.875442 4719 scope.go:117] "RemoveContainer" containerID="814739b420072df825f9394388cfccd76fa74cfe6e483208a287e6ab49f535b6" Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.876996 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4fln2"] Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.892035 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4fln2"] Dec 15 13:12:31 crc kubenswrapper[4719]: I1215 13:12:31.913638 4719 scope.go:117] "RemoveContainer" containerID="06ece604772c2accaf07157da0c3ab7916b3e1c7d2ca20fbc7e77809e725cc61" Dec 15 13:12:33 crc kubenswrapper[4719]: I1215 13:12:33.489807 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" path="/var/lib/kubelet/pods/cab1f22d-ecb1-470e-8b10-61b4b0f8817d/volumes" Dec 15 13:14:51 crc kubenswrapper[4719]: I1215 13:14:51.356448 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:14:51 crc kubenswrapper[4719]: I1215 13:14:51.357008 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.169458 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9"] Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170530 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170549 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170567 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerName="extract-content" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170576 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerName="extract-content" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170588 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351aab48-2379-4481-8752-82bd22525194" containerName="extract-content" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170595 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="351aab48-2379-4481-8752-82bd22525194" containerName="extract-content" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170607 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351aab48-2379-4481-8752-82bd22525194" containerName="extract-utilities" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170615 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="351aab48-2379-4481-8752-82bd22525194" containerName="extract-utilities" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170631 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerName="extract-content" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170641 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerName="extract-content" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170653 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351aab48-2379-4481-8752-82bd22525194" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170659 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="351aab48-2379-4481-8752-82bd22525194" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170673 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerName="extract-utilities" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170683 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerName="extract-utilities" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170702 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerName="extract-utilities" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170710 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerName="extract-utilities" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170722 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerName="extract-content" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170729 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerName="extract-content" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170740 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170747 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170768 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170778 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: E1215 13:15:00.170792 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerName="extract-utilities" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.170799 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerName="extract-utilities" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.171032 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="cab1f22d-ecb1-470e-8b10-61b4b0f8817d" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.171047 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccefa2e7-7772-4804-977b-f6600bbfe02d" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.171064 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e283d28-1b64-4b7f-aac2-b3ad186c6fcb" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.171082 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="351aab48-2379-4481-8752-82bd22525194" containerName="registry-server" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.173906 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.178099 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.178525 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.181219 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9"] Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.324436 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-secret-volume\") pod \"collect-profiles-29430075-bksk9\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.324756 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-config-volume\") pod \"collect-profiles-29430075-bksk9\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.324846 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzxxv\" (UniqueName: \"kubernetes.io/projected/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-kube-api-access-hzxxv\") pod \"collect-profiles-29430075-bksk9\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.426242 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-config-volume\") pod \"collect-profiles-29430075-bksk9\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.426399 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzxxv\" (UniqueName: \"kubernetes.io/projected/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-kube-api-access-hzxxv\") pod \"collect-profiles-29430075-bksk9\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.426478 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-secret-volume\") pod \"collect-profiles-29430075-bksk9\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.427366 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-config-volume\") pod \"collect-profiles-29430075-bksk9\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.433294 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-secret-volume\") pod \"collect-profiles-29430075-bksk9\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.447529 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzxxv\" (UniqueName: \"kubernetes.io/projected/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-kube-api-access-hzxxv\") pod \"collect-profiles-29430075-bksk9\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:00 crc kubenswrapper[4719]: I1215 13:15:00.507097 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:01 crc kubenswrapper[4719]: I1215 13:15:01.033745 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9"] Dec 15 13:15:01 crc kubenswrapper[4719]: I1215 13:15:01.152670 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" event={"ID":"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4","Type":"ContainerStarted","Data":"5fc5e75f5dc634c56d8dd0a9e067154c7c3bcac4c82672a40856d906ce386d1a"} Dec 15 13:15:02 crc kubenswrapper[4719]: I1215 13:15:02.162715 4719 generic.go:334] "Generic (PLEG): container finished" podID="e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4" containerID="26b7f76f465c0fc37d5e7739ed1e94546c161573e6e1d66fafb479662f614948" exitCode=0 Dec 15 13:15:02 crc kubenswrapper[4719]: I1215 13:15:02.162766 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" event={"ID":"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4","Type":"ContainerDied","Data":"26b7f76f465c0fc37d5e7739ed1e94546c161573e6e1d66fafb479662f614948"} Dec 15 13:15:03 crc kubenswrapper[4719]: I1215 13:15:03.890838 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.054988 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-secret-volume\") pod \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.055082 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-config-volume\") pod \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.055380 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzxxv\" (UniqueName: \"kubernetes.io/projected/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-kube-api-access-hzxxv\") pod \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\" (UID: \"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4\") " Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.056392 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-config-volume" (OuterVolumeSpecName: "config-volume") pod "e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4" (UID: "e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.066388 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4" (UID: "e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.068123 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-kube-api-access-hzxxv" (OuterVolumeSpecName: "kube-api-access-hzxxv") pod "e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4" (UID: "e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4"). InnerVolumeSpecName "kube-api-access-hzxxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.158417 4719 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.158455 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzxxv\" (UniqueName: \"kubernetes.io/projected/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-kube-api-access-hzxxv\") on node \"crc\" DevicePath \"\"" Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.158470 4719 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.180204 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" event={"ID":"e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4","Type":"ContainerDied","Data":"5fc5e75f5dc634c56d8dd0a9e067154c7c3bcac4c82672a40856d906ce386d1a"} Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.180262 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430075-bksk9" Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.180262 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fc5e75f5dc634c56d8dd0a9e067154c7c3bcac4c82672a40856d906ce386d1a" Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.976118 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb"] Dec 15 13:15:04 crc kubenswrapper[4719]: I1215 13:15:04.987516 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430030-vc5hb"] Dec 15 13:15:05 crc kubenswrapper[4719]: I1215 13:15:05.486645 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5" path="/var/lib/kubelet/pods/e36e55e0-a12e-4c1d-a5bd-987c29b6b6a5/volumes" Dec 15 13:15:21 crc kubenswrapper[4719]: I1215 13:15:21.356547 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:15:21 crc kubenswrapper[4719]: I1215 13:15:21.357085 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:15:35 crc kubenswrapper[4719]: I1215 13:15:35.176801 4719 scope.go:117] "RemoveContainer" containerID="97aaa509878268d2bb03c24a14b7861cbc39f40b96f5f043bf9e693599cba0c5" Dec 15 13:15:51 crc kubenswrapper[4719]: I1215 13:15:51.357142 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:15:51 crc kubenswrapper[4719]: I1215 13:15:51.357771 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:15:51 crc kubenswrapper[4719]: I1215 13:15:51.357825 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 13:15:51 crc kubenswrapper[4719]: I1215 13:15:51.358635 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3af3a7a5155192d69dc971d21bcb5ef1e92363efaa38759a082b973f30a65031"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 13:15:51 crc kubenswrapper[4719]: I1215 13:15:51.358695 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://3af3a7a5155192d69dc971d21bcb5ef1e92363efaa38759a082b973f30a65031" gracePeriod=600 Dec 15 13:15:51 crc kubenswrapper[4719]: I1215 13:15:51.580630 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="3af3a7a5155192d69dc971d21bcb5ef1e92363efaa38759a082b973f30a65031" exitCode=0 Dec 15 13:15:51 crc kubenswrapper[4719]: I1215 13:15:51.580672 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"3af3a7a5155192d69dc971d21bcb5ef1e92363efaa38759a082b973f30a65031"} Dec 15 13:15:51 crc kubenswrapper[4719]: I1215 13:15:51.580709 4719 scope.go:117] "RemoveContainer" containerID="9f2cde5eed89700ba61e3065783b8036420f722957530211747d74d32fa0df8d" Dec 15 13:15:52 crc kubenswrapper[4719]: I1215 13:15:52.594102 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013"} Dec 15 13:17:51 crc kubenswrapper[4719]: I1215 13:17:51.357768 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:17:51 crc kubenswrapper[4719]: I1215 13:17:51.358514 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:18:21 crc kubenswrapper[4719]: I1215 13:18:21.356974 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:18:21 crc kubenswrapper[4719]: I1215 13:18:21.357632 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:18:51 crc kubenswrapper[4719]: I1215 13:18:51.356898 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:18:51 crc kubenswrapper[4719]: I1215 13:18:51.357449 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:18:51 crc kubenswrapper[4719]: I1215 13:18:51.357498 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 13:18:51 crc kubenswrapper[4719]: I1215 13:18:51.358235 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 13:18:51 crc kubenswrapper[4719]: I1215 13:18:51.358386 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" gracePeriod=600 Dec 15 13:18:51 crc kubenswrapper[4719]: E1215 13:18:51.494495 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:18:52 crc kubenswrapper[4719]: I1215 13:18:52.172189 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" exitCode=0 Dec 15 13:18:52 crc kubenswrapper[4719]: I1215 13:18:52.172240 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013"} Dec 15 13:18:52 crc kubenswrapper[4719]: I1215 13:18:52.172278 4719 scope.go:117] "RemoveContainer" containerID="3af3a7a5155192d69dc971d21bcb5ef1e92363efaa38759a082b973f30a65031" Dec 15 13:18:52 crc kubenswrapper[4719]: I1215 13:18:52.173003 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:18:52 crc kubenswrapper[4719]: E1215 13:18:52.173448 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:19:04 crc kubenswrapper[4719]: I1215 13:19:04.476183 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:19:04 crc kubenswrapper[4719]: E1215 13:19:04.477087 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:19:16 crc kubenswrapper[4719]: I1215 13:19:16.476974 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:19:16 crc kubenswrapper[4719]: E1215 13:19:16.477783 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:19:28 crc kubenswrapper[4719]: I1215 13:19:28.476440 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:19:28 crc kubenswrapper[4719]: E1215 13:19:28.477269 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:19:39 crc kubenswrapper[4719]: I1215 13:19:39.486063 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:19:39 crc kubenswrapper[4719]: E1215 13:19:39.486909 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:19:51 crc kubenswrapper[4719]: I1215 13:19:51.476690 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:19:51 crc kubenswrapper[4719]: E1215 13:19:51.477494 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:20:02 crc kubenswrapper[4719]: I1215 13:20:02.476067 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:20:02 crc kubenswrapper[4719]: E1215 13:20:02.476886 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:20:13 crc kubenswrapper[4719]: I1215 13:20:13.476302 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:20:13 crc kubenswrapper[4719]: E1215 13:20:13.477153 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:20:28 crc kubenswrapper[4719]: I1215 13:20:28.475760 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:20:28 crc kubenswrapper[4719]: E1215 13:20:28.476427 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:20:41 crc kubenswrapper[4719]: I1215 13:20:41.479811 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:20:41 crc kubenswrapper[4719]: E1215 13:20:41.480418 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:20:54 crc kubenswrapper[4719]: I1215 13:20:54.475776 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:20:54 crc kubenswrapper[4719]: E1215 13:20:54.476571 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:21:05 crc kubenswrapper[4719]: I1215 13:21:05.477802 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:21:05 crc kubenswrapper[4719]: E1215 13:21:05.478552 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:21:19 crc kubenswrapper[4719]: I1215 13:21:19.522227 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:21:19 crc kubenswrapper[4719]: E1215 13:21:19.527101 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:21:33 crc kubenswrapper[4719]: I1215 13:21:33.476748 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:21:33 crc kubenswrapper[4719]: E1215 13:21:33.477454 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:21:47 crc kubenswrapper[4719]: I1215 13:21:47.477433 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:21:47 crc kubenswrapper[4719]: E1215 13:21:47.479256 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.734482 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bfpx4"] Dec 15 13:21:52 crc kubenswrapper[4719]: E1215 13:21:52.735413 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4" containerName="collect-profiles" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.735426 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4" containerName="collect-profiles" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.735631 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6d8abb2-71d5-47c1-af7a-df2c4b1c0ba4" containerName="collect-profiles" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.736949 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.764902 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfpx4"] Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.844071 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-catalog-content\") pod \"redhat-operators-bfpx4\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.844228 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chx4v\" (UniqueName: \"kubernetes.io/projected/b5fb2110-3b92-4e2b-b563-000e80fb506c-kube-api-access-chx4v\") pod \"redhat-operators-bfpx4\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.844283 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-utilities\") pod \"redhat-operators-bfpx4\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.945391 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-catalog-content\") pod \"redhat-operators-bfpx4\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.945839 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chx4v\" (UniqueName: \"kubernetes.io/projected/b5fb2110-3b92-4e2b-b563-000e80fb506c-kube-api-access-chx4v\") pod \"redhat-operators-bfpx4\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.945988 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-utilities\") pod \"redhat-operators-bfpx4\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.948053 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-catalog-content\") pod \"redhat-operators-bfpx4\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.948073 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-utilities\") pod \"redhat-operators-bfpx4\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:52 crc kubenswrapper[4719]: I1215 13:21:52.968328 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chx4v\" (UniqueName: \"kubernetes.io/projected/b5fb2110-3b92-4e2b-b563-000e80fb506c-kube-api-access-chx4v\") pod \"redhat-operators-bfpx4\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:53 crc kubenswrapper[4719]: I1215 13:21:53.113597 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:21:55 crc kubenswrapper[4719]: I1215 13:21:53.633796 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfpx4"] Dec 15 13:21:55 crc kubenswrapper[4719]: I1215 13:21:54.440930 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfpx4" event={"ID":"b5fb2110-3b92-4e2b-b563-000e80fb506c","Type":"ContainerStarted","Data":"af79ca4bb835a3b92b97ac326221d2cb584f7862982ac3df0166a192dee898db"} Dec 15 13:21:55 crc kubenswrapper[4719]: I1215 13:21:55.452932 4719 generic.go:334] "Generic (PLEG): container finished" podID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerID="067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a" exitCode=0 Dec 15 13:21:55 crc kubenswrapper[4719]: I1215 13:21:55.453125 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfpx4" event={"ID":"b5fb2110-3b92-4e2b-b563-000e80fb506c","Type":"ContainerDied","Data":"067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a"} Dec 15 13:21:55 crc kubenswrapper[4719]: I1215 13:21:55.456404 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 13:21:57 crc kubenswrapper[4719]: I1215 13:21:57.469155 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfpx4" event={"ID":"b5fb2110-3b92-4e2b-b563-000e80fb506c","Type":"ContainerStarted","Data":"6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c"} Dec 15 13:22:01 crc kubenswrapper[4719]: I1215 13:22:01.476633 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:22:01 crc kubenswrapper[4719]: E1215 13:22:01.477578 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:22:01 crc kubenswrapper[4719]: I1215 13:22:01.520756 4719 generic.go:334] "Generic (PLEG): container finished" podID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerID="6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c" exitCode=0 Dec 15 13:22:01 crc kubenswrapper[4719]: I1215 13:22:01.520817 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfpx4" event={"ID":"b5fb2110-3b92-4e2b-b563-000e80fb506c","Type":"ContainerDied","Data":"6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c"} Dec 15 13:22:03 crc kubenswrapper[4719]: I1215 13:22:03.536544 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfpx4" event={"ID":"b5fb2110-3b92-4e2b-b563-000e80fb506c","Type":"ContainerStarted","Data":"d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684"} Dec 15 13:22:03 crc kubenswrapper[4719]: I1215 13:22:03.567791 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bfpx4" podStartSLOduration=4.245287567 podStartE2EDuration="11.56776601s" podCreationTimestamp="2025-12-15 13:21:52 +0000 UTC" firstStartedPulling="2025-12-15 13:21:55.456171036 +0000 UTC m=+3876.398464066" lastFinishedPulling="2025-12-15 13:22:02.778649459 +0000 UTC m=+3883.720942509" observedRunningTime="2025-12-15 13:22:03.554108782 +0000 UTC m=+3884.496401832" watchObservedRunningTime="2025-12-15 13:22:03.56776601 +0000 UTC m=+3884.510059040" Dec 15 13:22:13 crc kubenswrapper[4719]: I1215 13:22:13.114423 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:22:13 crc kubenswrapper[4719]: I1215 13:22:13.115040 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:22:13 crc kubenswrapper[4719]: I1215 13:22:13.163113 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:22:13 crc kubenswrapper[4719]: I1215 13:22:13.477643 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:22:13 crc kubenswrapper[4719]: E1215 13:22:13.478199 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:22:13 crc kubenswrapper[4719]: I1215 13:22:13.664418 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:22:13 crc kubenswrapper[4719]: I1215 13:22:13.707724 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfpx4"] Dec 15 13:22:15 crc kubenswrapper[4719]: I1215 13:22:15.634508 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bfpx4" podUID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerName="registry-server" containerID="cri-o://d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684" gracePeriod=2 Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.220146 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.330643 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chx4v\" (UniqueName: \"kubernetes.io/projected/b5fb2110-3b92-4e2b-b563-000e80fb506c-kube-api-access-chx4v\") pod \"b5fb2110-3b92-4e2b-b563-000e80fb506c\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.331072 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-utilities\") pod \"b5fb2110-3b92-4e2b-b563-000e80fb506c\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.331280 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-catalog-content\") pod \"b5fb2110-3b92-4e2b-b563-000e80fb506c\" (UID: \"b5fb2110-3b92-4e2b-b563-000e80fb506c\") " Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.332568 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-utilities" (OuterVolumeSpecName: "utilities") pod "b5fb2110-3b92-4e2b-b563-000e80fb506c" (UID: "b5fb2110-3b92-4e2b-b563-000e80fb506c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.344289 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5fb2110-3b92-4e2b-b563-000e80fb506c-kube-api-access-chx4v" (OuterVolumeSpecName: "kube-api-access-chx4v") pod "b5fb2110-3b92-4e2b-b563-000e80fb506c" (UID: "b5fb2110-3b92-4e2b-b563-000e80fb506c"). InnerVolumeSpecName "kube-api-access-chx4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.434409 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chx4v\" (UniqueName: \"kubernetes.io/projected/b5fb2110-3b92-4e2b-b563-000e80fb506c-kube-api-access-chx4v\") on node \"crc\" DevicePath \"\"" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.434454 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.466215 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5fb2110-3b92-4e2b-b563-000e80fb506c" (UID: "b5fb2110-3b92-4e2b-b563-000e80fb506c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.536142 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fb2110-3b92-4e2b-b563-000e80fb506c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.644792 4719 generic.go:334] "Generic (PLEG): container finished" podID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerID="d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684" exitCode=0 Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.644867 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfpx4" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.644843 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfpx4" event={"ID":"b5fb2110-3b92-4e2b-b563-000e80fb506c","Type":"ContainerDied","Data":"d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684"} Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.644994 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfpx4" event={"ID":"b5fb2110-3b92-4e2b-b563-000e80fb506c","Type":"ContainerDied","Data":"af79ca4bb835a3b92b97ac326221d2cb584f7862982ac3df0166a192dee898db"} Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.645019 4719 scope.go:117] "RemoveContainer" containerID="d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.668363 4719 scope.go:117] "RemoveContainer" containerID="6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.688106 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfpx4"] Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.694337 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bfpx4"] Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.700612 4719 scope.go:117] "RemoveContainer" containerID="067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.739784 4719 scope.go:117] "RemoveContainer" containerID="d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684" Dec 15 13:22:16 crc kubenswrapper[4719]: E1215 13:22:16.740357 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684\": container with ID starting with d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684 not found: ID does not exist" containerID="d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.740389 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684"} err="failed to get container status \"d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684\": rpc error: code = NotFound desc = could not find container \"d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684\": container with ID starting with d2fb441379863162d581356687b3d3ea7775e0f97e60c242d30e7683c2067684 not found: ID does not exist" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.740418 4719 scope.go:117] "RemoveContainer" containerID="6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c" Dec 15 13:22:16 crc kubenswrapper[4719]: E1215 13:22:16.740752 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c\": container with ID starting with 6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c not found: ID does not exist" containerID="6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.740778 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c"} err="failed to get container status \"6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c\": rpc error: code = NotFound desc = could not find container \"6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c\": container with ID starting with 6a886a648a544059ec6aeeb5a412bc515de1f3f39ebfc2c114586bcc5734001c not found: ID does not exist" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.740797 4719 scope.go:117] "RemoveContainer" containerID="067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a" Dec 15 13:22:16 crc kubenswrapper[4719]: E1215 13:22:16.741177 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a\": container with ID starting with 067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a not found: ID does not exist" containerID="067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a" Dec 15 13:22:16 crc kubenswrapper[4719]: I1215 13:22:16.741240 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a"} err="failed to get container status \"067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a\": rpc error: code = NotFound desc = could not find container \"067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a\": container with ID starting with 067089154e2a885e5a208eef708eff260c4a801b93ebeec5b8257a471c64c68a not found: ID does not exist" Dec 15 13:22:17 crc kubenswrapper[4719]: I1215 13:22:17.485305 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5fb2110-3b92-4e2b-b563-000e80fb506c" path="/var/lib/kubelet/pods/b5fb2110-3b92-4e2b-b563-000e80fb506c/volumes" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.296604 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x7t7p"] Dec 15 13:22:28 crc kubenswrapper[4719]: E1215 13:22:28.297713 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerName="registry-server" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.297731 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerName="registry-server" Dec 15 13:22:28 crc kubenswrapper[4719]: E1215 13:22:28.297745 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerName="extract-content" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.297753 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerName="extract-content" Dec 15 13:22:28 crc kubenswrapper[4719]: E1215 13:22:28.297782 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerName="extract-utilities" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.297791 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerName="extract-utilities" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.298069 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5fb2110-3b92-4e2b-b563-000e80fb506c" containerName="registry-server" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.299709 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.312830 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7t7p"] Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.364262 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-catalog-content\") pod \"redhat-marketplace-x7t7p\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.364460 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwvwn\" (UniqueName: \"kubernetes.io/projected/8448a1db-1551-4000-98dc-83c55d62b470-kube-api-access-wwvwn\") pod \"redhat-marketplace-x7t7p\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.364645 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-utilities\") pod \"redhat-marketplace-x7t7p\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.466416 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-catalog-content\") pod \"redhat-marketplace-x7t7p\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.466513 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwvwn\" (UniqueName: \"kubernetes.io/projected/8448a1db-1551-4000-98dc-83c55d62b470-kube-api-access-wwvwn\") pod \"redhat-marketplace-x7t7p\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.466548 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-utilities\") pod \"redhat-marketplace-x7t7p\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.467045 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-catalog-content\") pod \"redhat-marketplace-x7t7p\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.467085 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-utilities\") pod \"redhat-marketplace-x7t7p\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.476654 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:22:28 crc kubenswrapper[4719]: E1215 13:22:28.476948 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.503013 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwvwn\" (UniqueName: \"kubernetes.io/projected/8448a1db-1551-4000-98dc-83c55d62b470-kube-api-access-wwvwn\") pod \"redhat-marketplace-x7t7p\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:28 crc kubenswrapper[4719]: I1215 13:22:28.622824 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:29 crc kubenswrapper[4719]: I1215 13:22:29.248909 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7t7p"] Dec 15 13:22:29 crc kubenswrapper[4719]: I1215 13:22:29.784822 4719 generic.go:334] "Generic (PLEG): container finished" podID="8448a1db-1551-4000-98dc-83c55d62b470" containerID="0eaaec9eebf67d546be4752deb6a35e7e47436669b339738a957132214e8aa65" exitCode=0 Dec 15 13:22:29 crc kubenswrapper[4719]: I1215 13:22:29.785027 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7t7p" event={"ID":"8448a1db-1551-4000-98dc-83c55d62b470","Type":"ContainerDied","Data":"0eaaec9eebf67d546be4752deb6a35e7e47436669b339738a957132214e8aa65"} Dec 15 13:22:29 crc kubenswrapper[4719]: I1215 13:22:29.785203 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7t7p" event={"ID":"8448a1db-1551-4000-98dc-83c55d62b470","Type":"ContainerStarted","Data":"984cdd082d9b56362d89432558038a0490c866575911728cf457f711942a2fb7"} Dec 15 13:22:30 crc kubenswrapper[4719]: I1215 13:22:30.802622 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7t7p" event={"ID":"8448a1db-1551-4000-98dc-83c55d62b470","Type":"ContainerStarted","Data":"703decea31704be0f92899636f16b5e21cbfe81454ceea46f5f5409642ed1677"} Dec 15 13:22:31 crc kubenswrapper[4719]: I1215 13:22:31.812029 4719 generic.go:334] "Generic (PLEG): container finished" podID="8448a1db-1551-4000-98dc-83c55d62b470" containerID="703decea31704be0f92899636f16b5e21cbfe81454ceea46f5f5409642ed1677" exitCode=0 Dec 15 13:22:31 crc kubenswrapper[4719]: I1215 13:22:31.812130 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7t7p" event={"ID":"8448a1db-1551-4000-98dc-83c55d62b470","Type":"ContainerDied","Data":"703decea31704be0f92899636f16b5e21cbfe81454ceea46f5f5409642ed1677"} Dec 15 13:22:33 crc kubenswrapper[4719]: I1215 13:22:33.832042 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7t7p" event={"ID":"8448a1db-1551-4000-98dc-83c55d62b470","Type":"ContainerStarted","Data":"43b736c7bca4b2b82703c47661680cb436401f7e4025f046314ded426d7d7cd6"} Dec 15 13:22:33 crc kubenswrapper[4719]: I1215 13:22:33.862539 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x7t7p" podStartSLOduration=2.684995838 podStartE2EDuration="5.862501281s" podCreationTimestamp="2025-12-15 13:22:28 +0000 UTC" firstStartedPulling="2025-12-15 13:22:29.7880818 +0000 UTC m=+3910.730374830" lastFinishedPulling="2025-12-15 13:22:32.965587243 +0000 UTC m=+3913.907880273" observedRunningTime="2025-12-15 13:22:33.854150489 +0000 UTC m=+3914.796443539" watchObservedRunningTime="2025-12-15 13:22:33.862501281 +0000 UTC m=+3914.804794311" Dec 15 13:22:38 crc kubenswrapper[4719]: I1215 13:22:38.623000 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:38 crc kubenswrapper[4719]: I1215 13:22:38.623608 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:38 crc kubenswrapper[4719]: I1215 13:22:38.674941 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:38 crc kubenswrapper[4719]: I1215 13:22:38.913615 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:38 crc kubenswrapper[4719]: I1215 13:22:38.966953 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7t7p"] Dec 15 13:22:40 crc kubenswrapper[4719]: I1215 13:22:40.885609 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x7t7p" podUID="8448a1db-1551-4000-98dc-83c55d62b470" containerName="registry-server" containerID="cri-o://43b736c7bca4b2b82703c47661680cb436401f7e4025f046314ded426d7d7cd6" gracePeriod=2 Dec 15 13:22:41 crc kubenswrapper[4719]: I1215 13:22:41.895730 4719 generic.go:334] "Generic (PLEG): container finished" podID="8448a1db-1551-4000-98dc-83c55d62b470" containerID="43b736c7bca4b2b82703c47661680cb436401f7e4025f046314ded426d7d7cd6" exitCode=0 Dec 15 13:22:41 crc kubenswrapper[4719]: I1215 13:22:41.895811 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7t7p" event={"ID":"8448a1db-1551-4000-98dc-83c55d62b470","Type":"ContainerDied","Data":"43b736c7bca4b2b82703c47661680cb436401f7e4025f046314ded426d7d7cd6"} Dec 15 13:22:41 crc kubenswrapper[4719]: I1215 13:22:41.929095 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mwdv9"] Dec 15 13:22:41 crc kubenswrapper[4719]: I1215 13:22:41.931106 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:41 crc kubenswrapper[4719]: I1215 13:22:41.942660 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mwdv9"] Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.074547 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp45h\" (UniqueName: \"kubernetes.io/projected/b90ede87-bd98-4a33-bfa9-65547250c5b3-kube-api-access-cp45h\") pod \"certified-operators-mwdv9\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.074694 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-utilities\") pod \"certified-operators-mwdv9\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.074754 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-catalog-content\") pod \"certified-operators-mwdv9\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.176636 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-utilities\") pod \"certified-operators-mwdv9\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.176746 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-catalog-content\") pod \"certified-operators-mwdv9\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.176919 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp45h\" (UniqueName: \"kubernetes.io/projected/b90ede87-bd98-4a33-bfa9-65547250c5b3-kube-api-access-cp45h\") pod \"certified-operators-mwdv9\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.177194 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-utilities\") pod \"certified-operators-mwdv9\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.177412 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-catalog-content\") pod \"certified-operators-mwdv9\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.198303 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp45h\" (UniqueName: \"kubernetes.io/projected/b90ede87-bd98-4a33-bfa9-65547250c5b3-kube-api-access-cp45h\") pod \"certified-operators-mwdv9\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.255592 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.441236 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.593104 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-utilities\") pod \"8448a1db-1551-4000-98dc-83c55d62b470\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.593648 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-catalog-content\") pod \"8448a1db-1551-4000-98dc-83c55d62b470\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.593683 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwvwn\" (UniqueName: \"kubernetes.io/projected/8448a1db-1551-4000-98dc-83c55d62b470-kube-api-access-wwvwn\") pod \"8448a1db-1551-4000-98dc-83c55d62b470\" (UID: \"8448a1db-1551-4000-98dc-83c55d62b470\") " Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.598364 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-utilities" (OuterVolumeSpecName: "utilities") pod "8448a1db-1551-4000-98dc-83c55d62b470" (UID: "8448a1db-1551-4000-98dc-83c55d62b470"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.623170 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8448a1db-1551-4000-98dc-83c55d62b470" (UID: "8448a1db-1551-4000-98dc-83c55d62b470"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.641051 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8448a1db-1551-4000-98dc-83c55d62b470-kube-api-access-wwvwn" (OuterVolumeSpecName: "kube-api-access-wwvwn") pod "8448a1db-1551-4000-98dc-83c55d62b470" (UID: "8448a1db-1551-4000-98dc-83c55d62b470"). InnerVolumeSpecName "kube-api-access-wwvwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.698394 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.698432 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwvwn\" (UniqueName: \"kubernetes.io/projected/8448a1db-1551-4000-98dc-83c55d62b470-kube-api-access-wwvwn\") on node \"crc\" DevicePath \"\"" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.698444 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8448a1db-1551-4000-98dc-83c55d62b470-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.908708 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7t7p" event={"ID":"8448a1db-1551-4000-98dc-83c55d62b470","Type":"ContainerDied","Data":"984cdd082d9b56362d89432558038a0490c866575911728cf457f711942a2fb7"} Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.909048 4719 scope.go:117] "RemoveContainer" containerID="43b736c7bca4b2b82703c47661680cb436401f7e4025f046314ded426d7d7cd6" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.908760 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7t7p" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.953562 4719 scope.go:117] "RemoveContainer" containerID="703decea31704be0f92899636f16b5e21cbfe81454ceea46f5f5409642ed1677" Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.955316 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7t7p"] Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.978309 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7t7p"] Dec 15 13:22:42 crc kubenswrapper[4719]: I1215 13:22:42.984458 4719 scope.go:117] "RemoveContainer" containerID="0eaaec9eebf67d546be4752deb6a35e7e47436669b339738a957132214e8aa65" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.054752 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mwdv9"] Dec 15 13:22:43 crc kubenswrapper[4719]: W1215 13:22:43.079222 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb90ede87_bd98_4a33_bfa9_65547250c5b3.slice/crio-995c368c68cf689d44039c58303a845a4666cf0a1145a9194f5568d4e5f312dd WatchSource:0}: Error finding container 995c368c68cf689d44039c58303a845a4666cf0a1145a9194f5568d4e5f312dd: Status 404 returned error can't find the container with id 995c368c68cf689d44039c58303a845a4666cf0a1145a9194f5568d4e5f312dd Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.476320 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:22:43 crc kubenswrapper[4719]: E1215 13:22:43.476828 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.487502 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8448a1db-1551-4000-98dc-83c55d62b470" path="/var/lib/kubelet/pods/8448a1db-1551-4000-98dc-83c55d62b470/volumes" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.757958 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9wpbc"] Dec 15 13:22:43 crc kubenswrapper[4719]: E1215 13:22:43.759374 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8448a1db-1551-4000-98dc-83c55d62b470" containerName="registry-server" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.759396 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="8448a1db-1551-4000-98dc-83c55d62b470" containerName="registry-server" Dec 15 13:22:43 crc kubenswrapper[4719]: E1215 13:22:43.759439 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8448a1db-1551-4000-98dc-83c55d62b470" containerName="extract-content" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.759445 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="8448a1db-1551-4000-98dc-83c55d62b470" containerName="extract-content" Dec 15 13:22:43 crc kubenswrapper[4719]: E1215 13:22:43.759491 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8448a1db-1551-4000-98dc-83c55d62b470" containerName="extract-utilities" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.759498 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="8448a1db-1551-4000-98dc-83c55d62b470" containerName="extract-utilities" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.760046 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="8448a1db-1551-4000-98dc-83c55d62b470" containerName="registry-server" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.764089 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.773805 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wpbc"] Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.851370 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwpwz\" (UniqueName: \"kubernetes.io/projected/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-kube-api-access-jwpwz\") pod \"community-operators-9wpbc\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.851479 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-utilities\") pod \"community-operators-9wpbc\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.851528 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-catalog-content\") pod \"community-operators-9wpbc\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.917832 4719 generic.go:334] "Generic (PLEG): container finished" podID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerID="737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1" exitCode=0 Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.917907 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mwdv9" event={"ID":"b90ede87-bd98-4a33-bfa9-65547250c5b3","Type":"ContainerDied","Data":"737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1"} Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.917930 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mwdv9" event={"ID":"b90ede87-bd98-4a33-bfa9-65547250c5b3","Type":"ContainerStarted","Data":"995c368c68cf689d44039c58303a845a4666cf0a1145a9194f5568d4e5f312dd"} Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.953223 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwpwz\" (UniqueName: \"kubernetes.io/projected/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-kube-api-access-jwpwz\") pod \"community-operators-9wpbc\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.953560 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-utilities\") pod \"community-operators-9wpbc\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.954229 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-catalog-content\") pod \"community-operators-9wpbc\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.954553 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-utilities\") pod \"community-operators-9wpbc\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:43 crc kubenswrapper[4719]: I1215 13:22:43.954789 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-catalog-content\") pod \"community-operators-9wpbc\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:44 crc kubenswrapper[4719]: I1215 13:22:44.010359 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwpwz\" (UniqueName: \"kubernetes.io/projected/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-kube-api-access-jwpwz\") pod \"community-operators-9wpbc\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:44 crc kubenswrapper[4719]: I1215 13:22:44.110433 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:44 crc kubenswrapper[4719]: I1215 13:22:44.719737 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wpbc"] Dec 15 13:22:44 crc kubenswrapper[4719]: I1215 13:22:44.929451 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wpbc" event={"ID":"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402","Type":"ContainerStarted","Data":"84e3af80502dfc34cdc45c355c6dfbd5c55c9ba7efb0b0e5979c7d217e9b6d95"} Dec 15 13:22:45 crc kubenswrapper[4719]: I1215 13:22:45.939227 4719 generic.go:334] "Generic (PLEG): container finished" podID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerID="b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76" exitCode=0 Dec 15 13:22:45 crc kubenswrapper[4719]: I1215 13:22:45.939284 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wpbc" event={"ID":"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402","Type":"ContainerDied","Data":"b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76"} Dec 15 13:22:45 crc kubenswrapper[4719]: I1215 13:22:45.943015 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mwdv9" event={"ID":"b90ede87-bd98-4a33-bfa9-65547250c5b3","Type":"ContainerStarted","Data":"5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df"} Dec 15 13:22:47 crc kubenswrapper[4719]: I1215 13:22:47.963096 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wpbc" event={"ID":"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402","Type":"ContainerStarted","Data":"d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53"} Dec 15 13:22:47 crc kubenswrapper[4719]: I1215 13:22:47.969941 4719 generic.go:334] "Generic (PLEG): container finished" podID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerID="5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df" exitCode=0 Dec 15 13:22:47 crc kubenswrapper[4719]: I1215 13:22:47.969979 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mwdv9" event={"ID":"b90ede87-bd98-4a33-bfa9-65547250c5b3","Type":"ContainerDied","Data":"5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df"} Dec 15 13:22:50 crc kubenswrapper[4719]: I1215 13:22:49.999440 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mwdv9" event={"ID":"b90ede87-bd98-4a33-bfa9-65547250c5b3","Type":"ContainerStarted","Data":"3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26"} Dec 15 13:22:50 crc kubenswrapper[4719]: I1215 13:22:50.005403 4719 generic.go:334] "Generic (PLEG): container finished" podID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerID="d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53" exitCode=0 Dec 15 13:22:50 crc kubenswrapper[4719]: I1215 13:22:50.005445 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wpbc" event={"ID":"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402","Type":"ContainerDied","Data":"d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53"} Dec 15 13:22:50 crc kubenswrapper[4719]: I1215 13:22:50.033370 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mwdv9" podStartSLOduration=3.515310794 podStartE2EDuration="9.033347629s" podCreationTimestamp="2025-12-15 13:22:41 +0000 UTC" firstStartedPulling="2025-12-15 13:22:43.920536562 +0000 UTC m=+3924.862829592" lastFinishedPulling="2025-12-15 13:22:49.438573387 +0000 UTC m=+3930.380866427" observedRunningTime="2025-12-15 13:22:50.019785195 +0000 UTC m=+3930.962078225" watchObservedRunningTime="2025-12-15 13:22:50.033347629 +0000 UTC m=+3930.975640659" Dec 15 13:22:51 crc kubenswrapper[4719]: I1215 13:22:51.017134 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wpbc" event={"ID":"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402","Type":"ContainerStarted","Data":"925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a"} Dec 15 13:22:51 crc kubenswrapper[4719]: I1215 13:22:51.043519 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9wpbc" podStartSLOduration=3.419515147 podStartE2EDuration="8.043494334s" podCreationTimestamp="2025-12-15 13:22:43 +0000 UTC" firstStartedPulling="2025-12-15 13:22:45.941131681 +0000 UTC m=+3926.883424711" lastFinishedPulling="2025-12-15 13:22:50.565110868 +0000 UTC m=+3931.507403898" observedRunningTime="2025-12-15 13:22:51.039450637 +0000 UTC m=+3931.981743667" watchObservedRunningTime="2025-12-15 13:22:51.043494334 +0000 UTC m=+3931.985787364" Dec 15 13:22:52 crc kubenswrapper[4719]: I1215 13:22:52.257120 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:52 crc kubenswrapper[4719]: I1215 13:22:52.257472 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:22:53 crc kubenswrapper[4719]: I1215 13:22:53.308041 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mwdv9" podUID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerName="registry-server" probeResult="failure" output=< Dec 15 13:22:53 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 13:22:53 crc kubenswrapper[4719]: > Dec 15 13:22:54 crc kubenswrapper[4719]: I1215 13:22:54.111100 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:54 crc kubenswrapper[4719]: I1215 13:22:54.111486 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:54 crc kubenswrapper[4719]: I1215 13:22:54.161539 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:55 crc kubenswrapper[4719]: I1215 13:22:55.102084 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:56 crc kubenswrapper[4719]: I1215 13:22:56.116781 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wpbc"] Dec 15 13:22:56 crc kubenswrapper[4719]: I1215 13:22:56.476311 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:22:56 crc kubenswrapper[4719]: E1215 13:22:56.476669 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.068629 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9wpbc" podUID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerName="registry-server" containerID="cri-o://925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a" gracePeriod=2 Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.724756 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.835115 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-catalog-content\") pod \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.835174 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwpwz\" (UniqueName: \"kubernetes.io/projected/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-kube-api-access-jwpwz\") pod \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.835267 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-utilities\") pod \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\" (UID: \"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402\") " Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.836070 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-utilities" (OuterVolumeSpecName: "utilities") pod "7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" (UID: "7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.840277 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-kube-api-access-jwpwz" (OuterVolumeSpecName: "kube-api-access-jwpwz") pod "7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" (UID: "7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402"). InnerVolumeSpecName "kube-api-access-jwpwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.890651 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" (UID: "7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.938152 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.938192 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwpwz\" (UniqueName: \"kubernetes.io/projected/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-kube-api-access-jwpwz\") on node \"crc\" DevicePath \"\"" Dec 15 13:22:57 crc kubenswrapper[4719]: I1215 13:22:57.938208 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.079669 4719 generic.go:334] "Generic (PLEG): container finished" podID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerID="925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a" exitCode=0 Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.079771 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wpbc" Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.079794 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wpbc" event={"ID":"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402","Type":"ContainerDied","Data":"925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a"} Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.080164 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wpbc" event={"ID":"7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402","Type":"ContainerDied","Data":"84e3af80502dfc34cdc45c355c6dfbd5c55c9ba7efb0b0e5979c7d217e9b6d95"} Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.080197 4719 scope.go:117] "RemoveContainer" containerID="925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a" Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.104800 4719 scope.go:117] "RemoveContainer" containerID="d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53" Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.129914 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wpbc"] Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.135739 4719 scope.go:117] "RemoveContainer" containerID="b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76" Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.142943 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9wpbc"] Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.174924 4719 scope.go:117] "RemoveContainer" containerID="925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a" Dec 15 13:22:58 crc kubenswrapper[4719]: E1215 13:22:58.175594 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a\": container with ID starting with 925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a not found: ID does not exist" containerID="925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a" Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.175656 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a"} err="failed to get container status \"925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a\": rpc error: code = NotFound desc = could not find container \"925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a\": container with ID starting with 925388643f4344d03d8db5c6281c468337dfb614d7518b1f54c72a1381aa5e9a not found: ID does not exist" Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.175684 4719 scope.go:117] "RemoveContainer" containerID="d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53" Dec 15 13:22:58 crc kubenswrapper[4719]: E1215 13:22:58.176169 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53\": container with ID starting with d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53 not found: ID does not exist" containerID="d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53" Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.176199 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53"} err="failed to get container status \"d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53\": rpc error: code = NotFound desc = could not find container \"d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53\": container with ID starting with d1b2bcb7f8fa24a59f5ab8674c8b6626168ea94548752efaf81cd6c47465bc53 not found: ID does not exist" Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.176222 4719 scope.go:117] "RemoveContainer" containerID="b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76" Dec 15 13:22:58 crc kubenswrapper[4719]: E1215 13:22:58.176515 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76\": container with ID starting with b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76 not found: ID does not exist" containerID="b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76" Dec 15 13:22:58 crc kubenswrapper[4719]: I1215 13:22:58.176544 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76"} err="failed to get container status \"b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76\": rpc error: code = NotFound desc = could not find container \"b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76\": container with ID starting with b5ce42662c299e0900c87480a8825e241a908d3a3c3a14026a2f01d05d705f76 not found: ID does not exist" Dec 15 13:22:59 crc kubenswrapper[4719]: I1215 13:22:59.485144 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" path="/var/lib/kubelet/pods/7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402/volumes" Dec 15 13:23:02 crc kubenswrapper[4719]: I1215 13:23:02.315955 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:23:02 crc kubenswrapper[4719]: I1215 13:23:02.367500 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:23:03 crc kubenswrapper[4719]: I1215 13:23:03.915867 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mwdv9"] Dec 15 13:23:04 crc kubenswrapper[4719]: I1215 13:23:04.137708 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mwdv9" podUID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerName="registry-server" containerID="cri-o://3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26" gracePeriod=2 Dec 15 13:23:04 crc kubenswrapper[4719]: I1215 13:23:04.888289 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:23:04 crc kubenswrapper[4719]: I1215 13:23:04.978553 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-utilities\") pod \"b90ede87-bd98-4a33-bfa9-65547250c5b3\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " Dec 15 13:23:04 crc kubenswrapper[4719]: I1215 13:23:04.978616 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-catalog-content\") pod \"b90ede87-bd98-4a33-bfa9-65547250c5b3\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " Dec 15 13:23:04 crc kubenswrapper[4719]: I1215 13:23:04.978799 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp45h\" (UniqueName: \"kubernetes.io/projected/b90ede87-bd98-4a33-bfa9-65547250c5b3-kube-api-access-cp45h\") pod \"b90ede87-bd98-4a33-bfa9-65547250c5b3\" (UID: \"b90ede87-bd98-4a33-bfa9-65547250c5b3\") " Dec 15 13:23:04 crc kubenswrapper[4719]: I1215 13:23:04.979267 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-utilities" (OuterVolumeSpecName: "utilities") pod "b90ede87-bd98-4a33-bfa9-65547250c5b3" (UID: "b90ede87-bd98-4a33-bfa9-65547250c5b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:23:04 crc kubenswrapper[4719]: I1215 13:23:04.979399 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:23:04 crc kubenswrapper[4719]: I1215 13:23:04.986117 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b90ede87-bd98-4a33-bfa9-65547250c5b3-kube-api-access-cp45h" (OuterVolumeSpecName: "kube-api-access-cp45h") pod "b90ede87-bd98-4a33-bfa9-65547250c5b3" (UID: "b90ede87-bd98-4a33-bfa9-65547250c5b3"). InnerVolumeSpecName "kube-api-access-cp45h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.039382 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b90ede87-bd98-4a33-bfa9-65547250c5b3" (UID: "b90ede87-bd98-4a33-bfa9-65547250c5b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.080961 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b90ede87-bd98-4a33-bfa9-65547250c5b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.081010 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp45h\" (UniqueName: \"kubernetes.io/projected/b90ede87-bd98-4a33-bfa9-65547250c5b3-kube-api-access-cp45h\") on node \"crc\" DevicePath \"\"" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.148109 4719 generic.go:334] "Generic (PLEG): container finished" podID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerID="3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26" exitCode=0 Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.148150 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mwdv9" event={"ID":"b90ede87-bd98-4a33-bfa9-65547250c5b3","Type":"ContainerDied","Data":"3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26"} Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.148179 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mwdv9" event={"ID":"b90ede87-bd98-4a33-bfa9-65547250c5b3","Type":"ContainerDied","Data":"995c368c68cf689d44039c58303a845a4666cf0a1145a9194f5568d4e5f312dd"} Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.148192 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mwdv9" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.148199 4719 scope.go:117] "RemoveContainer" containerID="3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.170470 4719 scope.go:117] "RemoveContainer" containerID="5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.202066 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mwdv9"] Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.203481 4719 scope.go:117] "RemoveContainer" containerID="737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.216411 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mwdv9"] Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.249352 4719 scope.go:117] "RemoveContainer" containerID="3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26" Dec 15 13:23:05 crc kubenswrapper[4719]: E1215 13:23:05.249792 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26\": container with ID starting with 3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26 not found: ID does not exist" containerID="3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.249822 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26"} err="failed to get container status \"3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26\": rpc error: code = NotFound desc = could not find container \"3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26\": container with ID starting with 3c1de88b2da52dc8aa4e30a75c89d8f070bb5eb19a6e7e7d42458ffdc3a72f26 not found: ID does not exist" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.249843 4719 scope.go:117] "RemoveContainer" containerID="5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df" Dec 15 13:23:05 crc kubenswrapper[4719]: E1215 13:23:05.250127 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df\": container with ID starting with 5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df not found: ID does not exist" containerID="5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.250170 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df"} err="failed to get container status \"5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df\": rpc error: code = NotFound desc = could not find container \"5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df\": container with ID starting with 5dde2ca335ea7277a5792251e27766a7023864c7ec0ceaece6870a62b16f53df not found: ID does not exist" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.250212 4719 scope.go:117] "RemoveContainer" containerID="737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1" Dec 15 13:23:05 crc kubenswrapper[4719]: E1215 13:23:05.250480 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1\": container with ID starting with 737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1 not found: ID does not exist" containerID="737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.250506 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1"} err="failed to get container status \"737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1\": rpc error: code = NotFound desc = could not find container \"737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1\": container with ID starting with 737b7d39a4879ba832336904e4a4a8b36ccc1eed53d0e43f3254663e4327b1a1 not found: ID does not exist" Dec 15 13:23:05 crc kubenswrapper[4719]: I1215 13:23:05.488387 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b90ede87-bd98-4a33-bfa9-65547250c5b3" path="/var/lib/kubelet/pods/b90ede87-bd98-4a33-bfa9-65547250c5b3/volumes" Dec 15 13:23:07 crc kubenswrapper[4719]: I1215 13:23:07.476509 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:23:07 crc kubenswrapper[4719]: E1215 13:23:07.477024 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:23:18 crc kubenswrapper[4719]: I1215 13:23:18.478250 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:23:18 crc kubenswrapper[4719]: E1215 13:23:18.479235 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:23:30 crc kubenswrapper[4719]: I1215 13:23:30.476287 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:23:30 crc kubenswrapper[4719]: E1215 13:23:30.477094 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:23:44 crc kubenswrapper[4719]: I1215 13:23:44.477132 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:23:44 crc kubenswrapper[4719]: E1215 13:23:44.478728 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:23:59 crc kubenswrapper[4719]: I1215 13:23:59.485788 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:24:00 crc kubenswrapper[4719]: I1215 13:24:00.643740 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"8d1b8c4ddbdff78f1aa922c1736e29bb40070105e17cc7baa34e4be277e72197"} Dec 15 13:26:21 crc kubenswrapper[4719]: I1215 13:26:21.356354 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:26:21 crc kubenswrapper[4719]: I1215 13:26:21.357019 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:26:51 crc kubenswrapper[4719]: I1215 13:26:51.356412 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:26:51 crc kubenswrapper[4719]: I1215 13:26:51.357030 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:27:21 crc kubenswrapper[4719]: I1215 13:27:21.356842 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:27:21 crc kubenswrapper[4719]: I1215 13:27:21.357462 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:27:21 crc kubenswrapper[4719]: I1215 13:27:21.357503 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 13:27:21 crc kubenswrapper[4719]: I1215 13:27:21.358216 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d1b8c4ddbdff78f1aa922c1736e29bb40070105e17cc7baa34e4be277e72197"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 13:27:21 crc kubenswrapper[4719]: I1215 13:27:21.358273 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://8d1b8c4ddbdff78f1aa922c1736e29bb40070105e17cc7baa34e4be277e72197" gracePeriod=600 Dec 15 13:27:22 crc kubenswrapper[4719]: I1215 13:27:22.307387 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="8d1b8c4ddbdff78f1aa922c1736e29bb40070105e17cc7baa34e4be277e72197" exitCode=0 Dec 15 13:27:22 crc kubenswrapper[4719]: I1215 13:27:22.307482 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"8d1b8c4ddbdff78f1aa922c1736e29bb40070105e17cc7baa34e4be277e72197"} Dec 15 13:27:22 crc kubenswrapper[4719]: I1215 13:27:22.308833 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7"} Dec 15 13:27:22 crc kubenswrapper[4719]: I1215 13:27:22.308890 4719 scope.go:117] "RemoveContainer" containerID="300e95a6c79fdb199b6210b93686f71c7abbdcebd8e20b4f0dc5e0c32814f013" Dec 15 13:29:21 crc kubenswrapper[4719]: I1215 13:29:21.356689 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:29:21 crc kubenswrapper[4719]: I1215 13:29:21.357549 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:29:51 crc kubenswrapper[4719]: I1215 13:29:51.357411 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:29:51 crc kubenswrapper[4719]: I1215 13:29:51.358009 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.177311 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt"] Dec 15 13:30:00 crc kubenswrapper[4719]: E1215 13:30:00.178413 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerName="extract-content" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.178434 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerName="extract-content" Dec 15 13:30:00 crc kubenswrapper[4719]: E1215 13:30:00.178462 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerName="extract-utilities" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.178471 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerName="extract-utilities" Dec 15 13:30:00 crc kubenswrapper[4719]: E1215 13:30:00.178485 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerName="extract-content" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.178493 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerName="extract-content" Dec 15 13:30:00 crc kubenswrapper[4719]: E1215 13:30:00.178510 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerName="registry-server" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.178519 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerName="registry-server" Dec 15 13:30:00 crc kubenswrapper[4719]: E1215 13:30:00.178537 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerName="extract-utilities" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.178546 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerName="extract-utilities" Dec 15 13:30:00 crc kubenswrapper[4719]: E1215 13:30:00.178562 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerName="registry-server" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.178571 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerName="registry-server" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.178784 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90ede87-bd98-4a33-bfa9-65547250c5b3" containerName="registry-server" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.178810 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c4cc7ad-249e-43c9-8cf8-0a65e2dc7402" containerName="registry-server" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.179627 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.183848 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.184379 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.189203 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt"] Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.339489 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wjwc\" (UniqueName: \"kubernetes.io/projected/b563e079-7983-4e09-b9a2-63e84e2dd0bc-kube-api-access-2wjwc\") pod \"collect-profiles-29430090-8ghxt\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.339925 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b563e079-7983-4e09-b9a2-63e84e2dd0bc-secret-volume\") pod \"collect-profiles-29430090-8ghxt\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.340007 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b563e079-7983-4e09-b9a2-63e84e2dd0bc-config-volume\") pod \"collect-profiles-29430090-8ghxt\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.441625 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wjwc\" (UniqueName: \"kubernetes.io/projected/b563e079-7983-4e09-b9a2-63e84e2dd0bc-kube-api-access-2wjwc\") pod \"collect-profiles-29430090-8ghxt\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.441736 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b563e079-7983-4e09-b9a2-63e84e2dd0bc-secret-volume\") pod \"collect-profiles-29430090-8ghxt\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.441768 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b563e079-7983-4e09-b9a2-63e84e2dd0bc-config-volume\") pod \"collect-profiles-29430090-8ghxt\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.442818 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b563e079-7983-4e09-b9a2-63e84e2dd0bc-config-volume\") pod \"collect-profiles-29430090-8ghxt\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.455321 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b563e079-7983-4e09-b9a2-63e84e2dd0bc-secret-volume\") pod \"collect-profiles-29430090-8ghxt\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.458808 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wjwc\" (UniqueName: \"kubernetes.io/projected/b563e079-7983-4e09-b9a2-63e84e2dd0bc-kube-api-access-2wjwc\") pod \"collect-profiles-29430090-8ghxt\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:00 crc kubenswrapper[4719]: I1215 13:30:00.505223 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:01 crc kubenswrapper[4719]: I1215 13:30:01.051552 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt"] Dec 15 13:30:01 crc kubenswrapper[4719]: I1215 13:30:01.247925 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" event={"ID":"b563e079-7983-4e09-b9a2-63e84e2dd0bc","Type":"ContainerStarted","Data":"125e9862dff74b1b55d02f8ae9216ac0d182532a2fcbbde32e3c4d1fc48bd7b9"} Dec 15 13:30:01 crc kubenswrapper[4719]: I1215 13:30:01.248234 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" event={"ID":"b563e079-7983-4e09-b9a2-63e84e2dd0bc","Type":"ContainerStarted","Data":"5f3589244b4ce1dda4be5f9fcb4d9a925df3492b4aa6f58ce76cc73255e069e0"} Dec 15 13:30:01 crc kubenswrapper[4719]: I1215 13:30:01.267227 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" podStartSLOduration=1.267207622 podStartE2EDuration="1.267207622s" podCreationTimestamp="2025-12-15 13:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 13:30:01.264968382 +0000 UTC m=+4362.207261412" watchObservedRunningTime="2025-12-15 13:30:01.267207622 +0000 UTC m=+4362.209500652" Dec 15 13:30:02 crc kubenswrapper[4719]: I1215 13:30:02.263275 4719 generic.go:334] "Generic (PLEG): container finished" podID="b563e079-7983-4e09-b9a2-63e84e2dd0bc" containerID="125e9862dff74b1b55d02f8ae9216ac0d182532a2fcbbde32e3c4d1fc48bd7b9" exitCode=0 Dec 15 13:30:02 crc kubenswrapper[4719]: I1215 13:30:02.263324 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" event={"ID":"b563e079-7983-4e09-b9a2-63e84e2dd0bc","Type":"ContainerDied","Data":"125e9862dff74b1b55d02f8ae9216ac0d182532a2fcbbde32e3c4d1fc48bd7b9"} Dec 15 13:30:03 crc kubenswrapper[4719]: I1215 13:30:03.695728 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:03 crc kubenswrapper[4719]: I1215 13:30:03.807513 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b563e079-7983-4e09-b9a2-63e84e2dd0bc-secret-volume\") pod \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " Dec 15 13:30:03 crc kubenswrapper[4719]: I1215 13:30:03.807559 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b563e079-7983-4e09-b9a2-63e84e2dd0bc-config-volume\") pod \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " Dec 15 13:30:03 crc kubenswrapper[4719]: I1215 13:30:03.807683 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wjwc\" (UniqueName: \"kubernetes.io/projected/b563e079-7983-4e09-b9a2-63e84e2dd0bc-kube-api-access-2wjwc\") pod \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\" (UID: \"b563e079-7983-4e09-b9a2-63e84e2dd0bc\") " Dec 15 13:30:03 crc kubenswrapper[4719]: I1215 13:30:03.808275 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b563e079-7983-4e09-b9a2-63e84e2dd0bc-config-volume" (OuterVolumeSpecName: "config-volume") pod "b563e079-7983-4e09-b9a2-63e84e2dd0bc" (UID: "b563e079-7983-4e09-b9a2-63e84e2dd0bc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 13:30:03 crc kubenswrapper[4719]: I1215 13:30:03.813825 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b563e079-7983-4e09-b9a2-63e84e2dd0bc-kube-api-access-2wjwc" (OuterVolumeSpecName: "kube-api-access-2wjwc") pod "b563e079-7983-4e09-b9a2-63e84e2dd0bc" (UID: "b563e079-7983-4e09-b9a2-63e84e2dd0bc"). InnerVolumeSpecName "kube-api-access-2wjwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:30:03 crc kubenswrapper[4719]: I1215 13:30:03.821992 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b563e079-7983-4e09-b9a2-63e84e2dd0bc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b563e079-7983-4e09-b9a2-63e84e2dd0bc" (UID: "b563e079-7983-4e09-b9a2-63e84e2dd0bc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:30:03 crc kubenswrapper[4719]: I1215 13:30:03.909711 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wjwc\" (UniqueName: \"kubernetes.io/projected/b563e079-7983-4e09-b9a2-63e84e2dd0bc-kube-api-access-2wjwc\") on node \"crc\" DevicePath \"\"" Dec 15 13:30:03 crc kubenswrapper[4719]: I1215 13:30:03.909749 4719 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b563e079-7983-4e09-b9a2-63e84e2dd0bc-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 15 13:30:03 crc kubenswrapper[4719]: I1215 13:30:03.909762 4719 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b563e079-7983-4e09-b9a2-63e84e2dd0bc-config-volume\") on node \"crc\" DevicePath \"\"" Dec 15 13:30:04 crc kubenswrapper[4719]: I1215 13:30:04.281030 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" event={"ID":"b563e079-7983-4e09-b9a2-63e84e2dd0bc","Type":"ContainerDied","Data":"5f3589244b4ce1dda4be5f9fcb4d9a925df3492b4aa6f58ce76cc73255e069e0"} Dec 15 13:30:04 crc kubenswrapper[4719]: I1215 13:30:04.281087 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f3589244b4ce1dda4be5f9fcb4d9a925df3492b4aa6f58ce76cc73255e069e0" Dec 15 13:30:04 crc kubenswrapper[4719]: I1215 13:30:04.281180 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430090-8ghxt" Dec 15 13:30:04 crc kubenswrapper[4719]: I1215 13:30:04.340428 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q"] Dec 15 13:30:04 crc kubenswrapper[4719]: I1215 13:30:04.348358 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430045-v2g7q"] Dec 15 13:30:05 crc kubenswrapper[4719]: I1215 13:30:05.493091 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4537bedf-a3f6-4c61-b53f-2438beff6c75" path="/var/lib/kubelet/pods/4537bedf-a3f6-4c61-b53f-2438beff6c75/volumes" Dec 15 13:30:21 crc kubenswrapper[4719]: I1215 13:30:21.356585 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:30:21 crc kubenswrapper[4719]: I1215 13:30:21.357985 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:30:21 crc kubenswrapper[4719]: I1215 13:30:21.358057 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 13:30:21 crc kubenswrapper[4719]: I1215 13:30:21.358766 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 13:30:21 crc kubenswrapper[4719]: I1215 13:30:21.358839 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" gracePeriod=600 Dec 15 13:30:21 crc kubenswrapper[4719]: E1215 13:30:21.486603 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:30:22 crc kubenswrapper[4719]: I1215 13:30:22.434466 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" exitCode=0 Dec 15 13:30:22 crc kubenswrapper[4719]: I1215 13:30:22.434509 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7"} Dec 15 13:30:22 crc kubenswrapper[4719]: I1215 13:30:22.434562 4719 scope.go:117] "RemoveContainer" containerID="8d1b8c4ddbdff78f1aa922c1736e29bb40070105e17cc7baa34e4be277e72197" Dec 15 13:30:22 crc kubenswrapper[4719]: I1215 13:30:22.435361 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:30:22 crc kubenswrapper[4719]: E1215 13:30:22.435648 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:30:33 crc kubenswrapper[4719]: I1215 13:30:33.476749 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:30:33 crc kubenswrapper[4719]: E1215 13:30:33.477574 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:30:35 crc kubenswrapper[4719]: I1215 13:30:35.539997 4719 scope.go:117] "RemoveContainer" containerID="685de466f1d4bfacb9a74afddc5a3277dd1b20873ba53fe5bd954fbc2ba1f3c4" Dec 15 13:30:46 crc kubenswrapper[4719]: I1215 13:30:46.476156 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:30:46 crc kubenswrapper[4719]: E1215 13:30:46.476855 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:31:01 crc kubenswrapper[4719]: I1215 13:31:01.476141 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:31:01 crc kubenswrapper[4719]: E1215 13:31:01.477225 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:31:16 crc kubenswrapper[4719]: I1215 13:31:16.476246 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:31:16 crc kubenswrapper[4719]: E1215 13:31:16.477203 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:31:27 crc kubenswrapper[4719]: I1215 13:31:27.475886 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:31:27 crc kubenswrapper[4719]: E1215 13:31:27.476623 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:31:42 crc kubenswrapper[4719]: I1215 13:31:42.475841 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:31:42 crc kubenswrapper[4719]: E1215 13:31:42.476569 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:31:57 crc kubenswrapper[4719]: I1215 13:31:57.476431 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:31:57 crc kubenswrapper[4719]: E1215 13:31:57.477348 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:32:10 crc kubenswrapper[4719]: I1215 13:32:10.476610 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:32:10 crc kubenswrapper[4719]: E1215 13:32:10.477569 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.589244 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rdlv2"] Dec 15 13:32:23 crc kubenswrapper[4719]: E1215 13:32:23.591170 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b563e079-7983-4e09-b9a2-63e84e2dd0bc" containerName="collect-profiles" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.591191 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b563e079-7983-4e09-b9a2-63e84e2dd0bc" containerName="collect-profiles" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.591451 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="b563e079-7983-4e09-b9a2-63e84e2dd0bc" containerName="collect-profiles" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.593159 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.600248 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdlv2"] Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.645061 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtsv8\" (UniqueName: \"kubernetes.io/projected/a57e9044-d7cb-4f72-99e4-7def6534ce5c-kube-api-access-xtsv8\") pod \"redhat-operators-rdlv2\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.645104 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-catalog-content\") pod \"redhat-operators-rdlv2\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.645137 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-utilities\") pod \"redhat-operators-rdlv2\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.747226 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-utilities\") pod \"redhat-operators-rdlv2\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.747436 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtsv8\" (UniqueName: \"kubernetes.io/projected/a57e9044-d7cb-4f72-99e4-7def6534ce5c-kube-api-access-xtsv8\") pod \"redhat-operators-rdlv2\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.747464 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-catalog-content\") pod \"redhat-operators-rdlv2\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.747957 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-catalog-content\") pod \"redhat-operators-rdlv2\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.748094 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-utilities\") pod \"redhat-operators-rdlv2\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.772429 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtsv8\" (UniqueName: \"kubernetes.io/projected/a57e9044-d7cb-4f72-99e4-7def6534ce5c-kube-api-access-xtsv8\") pod \"redhat-operators-rdlv2\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:23 crc kubenswrapper[4719]: I1215 13:32:23.913455 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:24 crc kubenswrapper[4719]: I1215 13:32:24.476756 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:32:24 crc kubenswrapper[4719]: E1215 13:32:24.477219 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:32:24 crc kubenswrapper[4719]: I1215 13:32:24.783898 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdlv2"] Dec 15 13:32:25 crc kubenswrapper[4719]: I1215 13:32:25.472118 4719 generic.go:334] "Generic (PLEG): container finished" podID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerID="8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15" exitCode=0 Dec 15 13:32:25 crc kubenswrapper[4719]: I1215 13:32:25.472721 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdlv2" event={"ID":"a57e9044-d7cb-4f72-99e4-7def6534ce5c","Type":"ContainerDied","Data":"8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15"} Dec 15 13:32:25 crc kubenswrapper[4719]: I1215 13:32:25.472788 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdlv2" event={"ID":"a57e9044-d7cb-4f72-99e4-7def6534ce5c","Type":"ContainerStarted","Data":"b2e39cc61a72cb2ac784561081ed5c17aea8b101e23ac09a8bd9dde4d572d113"} Dec 15 13:32:25 crc kubenswrapper[4719]: I1215 13:32:25.475569 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 13:32:27 crc kubenswrapper[4719]: I1215 13:32:27.498249 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdlv2" event={"ID":"a57e9044-d7cb-4f72-99e4-7def6534ce5c","Type":"ContainerStarted","Data":"ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848"} Dec 15 13:32:30 crc kubenswrapper[4719]: I1215 13:32:30.527054 4719 generic.go:334] "Generic (PLEG): container finished" podID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerID="ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848" exitCode=0 Dec 15 13:32:30 crc kubenswrapper[4719]: I1215 13:32:30.527079 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdlv2" event={"ID":"a57e9044-d7cb-4f72-99e4-7def6534ce5c","Type":"ContainerDied","Data":"ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848"} Dec 15 13:32:31 crc kubenswrapper[4719]: I1215 13:32:31.537731 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdlv2" event={"ID":"a57e9044-d7cb-4f72-99e4-7def6534ce5c","Type":"ContainerStarted","Data":"3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4"} Dec 15 13:32:31 crc kubenswrapper[4719]: I1215 13:32:31.560585 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rdlv2" podStartSLOduration=2.882488966 podStartE2EDuration="8.560567115s" podCreationTimestamp="2025-12-15 13:32:23 +0000 UTC" firstStartedPulling="2025-12-15 13:32:25.475300257 +0000 UTC m=+4506.417593287" lastFinishedPulling="2025-12-15 13:32:31.153378406 +0000 UTC m=+4512.095671436" observedRunningTime="2025-12-15 13:32:31.560133021 +0000 UTC m=+4512.502426051" watchObservedRunningTime="2025-12-15 13:32:31.560567115 +0000 UTC m=+4512.502860145" Dec 15 13:32:33 crc kubenswrapper[4719]: I1215 13:32:33.914151 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:33 crc kubenswrapper[4719]: I1215 13:32:33.914215 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:34 crc kubenswrapper[4719]: I1215 13:32:34.960006 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rdlv2" podUID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerName="registry-server" probeResult="failure" output=< Dec 15 13:32:34 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 13:32:34 crc kubenswrapper[4719]: > Dec 15 13:32:36 crc kubenswrapper[4719]: I1215 13:32:36.476736 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:32:36 crc kubenswrapper[4719]: E1215 13:32:36.477261 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:32:43 crc kubenswrapper[4719]: I1215 13:32:43.960755 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:44 crc kubenswrapper[4719]: I1215 13:32:44.005013 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:44 crc kubenswrapper[4719]: I1215 13:32:44.203112 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdlv2"] Dec 15 13:32:45 crc kubenswrapper[4719]: I1215 13:32:45.644335 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rdlv2" podUID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerName="registry-server" containerID="cri-o://3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4" gracePeriod=2 Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.213335 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.382741 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtsv8\" (UniqueName: \"kubernetes.io/projected/a57e9044-d7cb-4f72-99e4-7def6534ce5c-kube-api-access-xtsv8\") pod \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.383176 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-utilities\") pod \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.383409 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-catalog-content\") pod \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\" (UID: \"a57e9044-d7cb-4f72-99e4-7def6534ce5c\") " Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.386382 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-utilities" (OuterVolumeSpecName: "utilities") pod "a57e9044-d7cb-4f72-99e4-7def6534ce5c" (UID: "a57e9044-d7cb-4f72-99e4-7def6534ce5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.391422 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a57e9044-d7cb-4f72-99e4-7def6534ce5c-kube-api-access-xtsv8" (OuterVolumeSpecName: "kube-api-access-xtsv8") pod "a57e9044-d7cb-4f72-99e4-7def6534ce5c" (UID: "a57e9044-d7cb-4f72-99e4-7def6534ce5c"). InnerVolumeSpecName "kube-api-access-xtsv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.485483 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtsv8\" (UniqueName: \"kubernetes.io/projected/a57e9044-d7cb-4f72-99e4-7def6534ce5c-kube-api-access-xtsv8\") on node \"crc\" DevicePath \"\"" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.485518 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.490574 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a57e9044-d7cb-4f72-99e4-7def6534ce5c" (UID: "a57e9044-d7cb-4f72-99e4-7def6534ce5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.587177 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57e9044-d7cb-4f72-99e4-7def6534ce5c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.653392 4719 generic.go:334] "Generic (PLEG): container finished" podID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerID="3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4" exitCode=0 Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.653443 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdlv2" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.654356 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdlv2" event={"ID":"a57e9044-d7cb-4f72-99e4-7def6534ce5c","Type":"ContainerDied","Data":"3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4"} Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.654451 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdlv2" event={"ID":"a57e9044-d7cb-4f72-99e4-7def6534ce5c","Type":"ContainerDied","Data":"b2e39cc61a72cb2ac784561081ed5c17aea8b101e23ac09a8bd9dde4d572d113"} Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.654532 4719 scope.go:117] "RemoveContainer" containerID="3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.682830 4719 scope.go:117] "RemoveContainer" containerID="ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.691613 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdlv2"] Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.703622 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rdlv2"] Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.704273 4719 scope.go:117] "RemoveContainer" containerID="8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.746119 4719 scope.go:117] "RemoveContainer" containerID="3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4" Dec 15 13:32:46 crc kubenswrapper[4719]: E1215 13:32:46.746763 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4\": container with ID starting with 3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4 not found: ID does not exist" containerID="3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.746797 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4"} err="failed to get container status \"3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4\": rpc error: code = NotFound desc = could not find container \"3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4\": container with ID starting with 3107081306cd05766bb3bb97c0bbf7561bdf93fa8f959a2a417bcaa8319f0eb4 not found: ID does not exist" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.746818 4719 scope.go:117] "RemoveContainer" containerID="ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848" Dec 15 13:32:46 crc kubenswrapper[4719]: E1215 13:32:46.748620 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848\": container with ID starting with ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848 not found: ID does not exist" containerID="ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.748648 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848"} err="failed to get container status \"ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848\": rpc error: code = NotFound desc = could not find container \"ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848\": container with ID starting with ea9724ecd1b7998035b52e3236e3023e4aaca9a2cad9c2e73e5b6ef49bdc8848 not found: ID does not exist" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.748665 4719 scope.go:117] "RemoveContainer" containerID="8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15" Dec 15 13:32:46 crc kubenswrapper[4719]: E1215 13:32:46.748923 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15\": container with ID starting with 8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15 not found: ID does not exist" containerID="8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15" Dec 15 13:32:46 crc kubenswrapper[4719]: I1215 13:32:46.748953 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15"} err="failed to get container status \"8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15\": rpc error: code = NotFound desc = could not find container \"8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15\": container with ID starting with 8be13969c2d345d82266e959e7d3f3e89f579514ff663d4b54d8fc688c7bcd15 not found: ID does not exist" Dec 15 13:32:47 crc kubenswrapper[4719]: I1215 13:32:47.486106 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" path="/var/lib/kubelet/pods/a57e9044-d7cb-4f72-99e4-7def6534ce5c/volumes" Dec 15 13:32:48 crc kubenswrapper[4719]: I1215 13:32:48.476028 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:32:48 crc kubenswrapper[4719]: E1215 13:32:48.476310 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:33:00 crc kubenswrapper[4719]: I1215 13:33:00.476129 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:33:00 crc kubenswrapper[4719]: E1215 13:33:00.476712 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:33:14 crc kubenswrapper[4719]: I1215 13:33:14.476233 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:33:14 crc kubenswrapper[4719]: E1215 13:33:14.477084 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:33:25 crc kubenswrapper[4719]: I1215 13:33:25.476905 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:33:25 crc kubenswrapper[4719]: E1215 13:33:25.477761 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:33:40 crc kubenswrapper[4719]: I1215 13:33:40.477014 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:33:40 crc kubenswrapper[4719]: E1215 13:33:40.478116 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:33:48 crc kubenswrapper[4719]: I1215 13:33:48.126039 4719 generic.go:334] "Generic (PLEG): container finished" podID="8b81902c-c979-4e03-a40f-cac54a63d6db" containerID="5ad12e1d824e57bb6f5d03028c8ed11cc6b0b0bebce7e6a040bceb64737fdda2" exitCode=0 Dec 15 13:33:48 crc kubenswrapper[4719]: I1215 13:33:48.126178 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8b81902c-c979-4e03-a40f-cac54a63d6db","Type":"ContainerDied","Data":"5ad12e1d824e57bb6f5d03028c8ed11cc6b0b0bebce7e6a040bceb64737fdda2"} Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.546964 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.654328 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-temporary\") pod \"8b81902c-c979-4e03-a40f-cac54a63d6db\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.654413 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"8b81902c-c979-4e03-a40f-cac54a63d6db\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.654443 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-config-data\") pod \"8b81902c-c979-4e03-a40f-cac54a63d6db\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.654470 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config-secret\") pod \"8b81902c-c979-4e03-a40f-cac54a63d6db\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.654492 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ssh-key\") pod \"8b81902c-c979-4e03-a40f-cac54a63d6db\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.654593 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvsjq\" (UniqueName: \"kubernetes.io/projected/8b81902c-c979-4e03-a40f-cac54a63d6db-kube-api-access-zvsjq\") pod \"8b81902c-c979-4e03-a40f-cac54a63d6db\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.654653 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config\") pod \"8b81902c-c979-4e03-a40f-cac54a63d6db\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.654711 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ca-certs\") pod \"8b81902c-c979-4e03-a40f-cac54a63d6db\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.654753 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-workdir\") pod \"8b81902c-c979-4e03-a40f-cac54a63d6db\" (UID: \"8b81902c-c979-4e03-a40f-cac54a63d6db\") " Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.654887 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "8b81902c-c979-4e03-a40f-cac54a63d6db" (UID: "8b81902c-c979-4e03-a40f-cac54a63d6db"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.655251 4719 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.655728 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-config-data" (OuterVolumeSpecName: "config-data") pod "8b81902c-c979-4e03-a40f-cac54a63d6db" (UID: "8b81902c-c979-4e03-a40f-cac54a63d6db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.659824 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b81902c-c979-4e03-a40f-cac54a63d6db-kube-api-access-zvsjq" (OuterVolumeSpecName: "kube-api-access-zvsjq") pod "8b81902c-c979-4e03-a40f-cac54a63d6db" (UID: "8b81902c-c979-4e03-a40f-cac54a63d6db"). InnerVolumeSpecName "kube-api-access-zvsjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.660139 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "8b81902c-c979-4e03-a40f-cac54a63d6db" (UID: "8b81902c-c979-4e03-a40f-cac54a63d6db"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.663955 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "8b81902c-c979-4e03-a40f-cac54a63d6db" (UID: "8b81902c-c979-4e03-a40f-cac54a63d6db"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.681712 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8b81902c-c979-4e03-a40f-cac54a63d6db" (UID: "8b81902c-c979-4e03-a40f-cac54a63d6db"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.682675 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8b81902c-c979-4e03-a40f-cac54a63d6db" (UID: "8b81902c-c979-4e03-a40f-cac54a63d6db"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.688602 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "8b81902c-c979-4e03-a40f-cac54a63d6db" (UID: "8b81902c-c979-4e03-a40f-cac54a63d6db"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.708961 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "8b81902c-c979-4e03-a40f-cac54a63d6db" (UID: "8b81902c-c979-4e03-a40f-cac54a63d6db"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.757728 4719 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.757762 4719 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-config-data\") on node \"crc\" DevicePath \"\"" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.757777 4719 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.757790 4719 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.757803 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvsjq\" (UniqueName: \"kubernetes.io/projected/8b81902c-c979-4e03-a40f-cac54a63d6db-kube-api-access-zvsjq\") on node \"crc\" DevicePath \"\"" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.757815 4719 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8b81902c-c979-4e03-a40f-cac54a63d6db-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.757826 4719 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8b81902c-c979-4e03-a40f-cac54a63d6db-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.757841 4719 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8b81902c-c979-4e03-a40f-cac54a63d6db-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.784519 4719 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 15 13:33:49 crc kubenswrapper[4719]: I1215 13:33:49.859480 4719 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 15 13:33:50 crc kubenswrapper[4719]: I1215 13:33:50.142876 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8b81902c-c979-4e03-a40f-cac54a63d6db","Type":"ContainerDied","Data":"ea5b6876af2f72484e40690c8757871aeaa90838994f480b0ee02c341864f907"} Dec 15 13:33:50 crc kubenswrapper[4719]: I1215 13:33:50.143218 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea5b6876af2f72484e40690c8757871aeaa90838994f480b0ee02c341864f907" Dec 15 13:33:50 crc kubenswrapper[4719]: I1215 13:33:50.142912 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.169387 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 15 13:33:54 crc kubenswrapper[4719]: E1215 13:33:54.170388 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerName="registry-server" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.170403 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerName="registry-server" Dec 15 13:33:54 crc kubenswrapper[4719]: E1215 13:33:54.170414 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b81902c-c979-4e03-a40f-cac54a63d6db" containerName="tempest-tests-tempest-tests-runner" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.170420 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b81902c-c979-4e03-a40f-cac54a63d6db" containerName="tempest-tests-tempest-tests-runner" Dec 15 13:33:54 crc kubenswrapper[4719]: E1215 13:33:54.170431 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerName="extract-content" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.170437 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerName="extract-content" Dec 15 13:33:54 crc kubenswrapper[4719]: E1215 13:33:54.170456 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerName="extract-utilities" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.170462 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerName="extract-utilities" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.170643 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b81902c-c979-4e03-a40f-cac54a63d6db" containerName="tempest-tests-tempest-tests-runner" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.170657 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="a57e9044-d7cb-4f72-99e4-7def6534ce5c" containerName="registry-server" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.171255 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.175361 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mlh57" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.240419 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.342737 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d347fbd6-e218-4153-be3f-f912f19123bb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.342791 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4zk2\" (UniqueName: \"kubernetes.io/projected/d347fbd6-e218-4153-be3f-f912f19123bb-kube-api-access-l4zk2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d347fbd6-e218-4153-be3f-f912f19123bb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.444674 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d347fbd6-e218-4153-be3f-f912f19123bb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.444751 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4zk2\" (UniqueName: \"kubernetes.io/projected/d347fbd6-e218-4153-be3f-f912f19123bb-kube-api-access-l4zk2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d347fbd6-e218-4153-be3f-f912f19123bb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.445520 4719 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d347fbd6-e218-4153-be3f-f912f19123bb\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.471903 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4zk2\" (UniqueName: \"kubernetes.io/projected/d347fbd6-e218-4153-be3f-f912f19123bb-kube-api-access-l4zk2\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d347fbd6-e218-4153-be3f-f912f19123bb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.473302 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"d347fbd6-e218-4153-be3f-f912f19123bb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.476913 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:33:54 crc kubenswrapper[4719]: E1215 13:33:54.477369 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.494427 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 15 13:33:54 crc kubenswrapper[4719]: I1215 13:33:54.933075 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 15 13:33:55 crc kubenswrapper[4719]: I1215 13:33:55.187277 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d347fbd6-e218-4153-be3f-f912f19123bb","Type":"ContainerStarted","Data":"230ea6b23f12b5117dac283a45f9fbefed7d037863dfc5302165b4b440d8a4c4"} Dec 15 13:33:56 crc kubenswrapper[4719]: I1215 13:33:56.196586 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"d347fbd6-e218-4153-be3f-f912f19123bb","Type":"ContainerStarted","Data":"8b5e580a58151d8c45ca01c50917ed45e6b869f5b3716560d23e8dfb9dfa88df"} Dec 15 13:33:56 crc kubenswrapper[4719]: I1215 13:33:56.212626 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.222842754 podStartE2EDuration="2.212605084s" podCreationTimestamp="2025-12-15 13:33:54 +0000 UTC" firstStartedPulling="2025-12-15 13:33:54.941003321 +0000 UTC m=+4595.883296361" lastFinishedPulling="2025-12-15 13:33:55.930765661 +0000 UTC m=+4596.873058691" observedRunningTime="2025-12-15 13:33:56.209259149 +0000 UTC m=+4597.151552189" watchObservedRunningTime="2025-12-15 13:33:56.212605084 +0000 UTC m=+4597.154898124" Dec 15 13:34:05 crc kubenswrapper[4719]: I1215 13:34:05.475974 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:34:05 crc kubenswrapper[4719]: E1215 13:34:05.476732 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:34:18 crc kubenswrapper[4719]: I1215 13:34:18.476717 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:34:18 crc kubenswrapper[4719]: E1215 13:34:18.478349 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.607059 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c8s97/must-gather-88wmh"] Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.609021 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/must-gather-88wmh" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.611701 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-c8s97"/"default-dockercfg-qsh4m" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.612090 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-c8s97"/"kube-root-ca.crt" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.620277 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-c8s97/must-gather-88wmh"] Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.623902 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-c8s97"/"openshift-service-ca.crt" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.766457 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcj72\" (UniqueName: \"kubernetes.io/projected/2397d720-99e1-4da0-93ff-aac39122fadf-kube-api-access-dcj72\") pod \"must-gather-88wmh\" (UID: \"2397d720-99e1-4da0-93ff-aac39122fadf\") " pod="openshift-must-gather-c8s97/must-gather-88wmh" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.766620 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2397d720-99e1-4da0-93ff-aac39122fadf-must-gather-output\") pod \"must-gather-88wmh\" (UID: \"2397d720-99e1-4da0-93ff-aac39122fadf\") " pod="openshift-must-gather-c8s97/must-gather-88wmh" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.867974 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2397d720-99e1-4da0-93ff-aac39122fadf-must-gather-output\") pod \"must-gather-88wmh\" (UID: \"2397d720-99e1-4da0-93ff-aac39122fadf\") " pod="openshift-must-gather-c8s97/must-gather-88wmh" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.868365 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcj72\" (UniqueName: \"kubernetes.io/projected/2397d720-99e1-4da0-93ff-aac39122fadf-kube-api-access-dcj72\") pod \"must-gather-88wmh\" (UID: \"2397d720-99e1-4da0-93ff-aac39122fadf\") " pod="openshift-must-gather-c8s97/must-gather-88wmh" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.868472 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2397d720-99e1-4da0-93ff-aac39122fadf-must-gather-output\") pod \"must-gather-88wmh\" (UID: \"2397d720-99e1-4da0-93ff-aac39122fadf\") " pod="openshift-must-gather-c8s97/must-gather-88wmh" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.898891 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcj72\" (UniqueName: \"kubernetes.io/projected/2397d720-99e1-4da0-93ff-aac39122fadf-kube-api-access-dcj72\") pod \"must-gather-88wmh\" (UID: \"2397d720-99e1-4da0-93ff-aac39122fadf\") " pod="openshift-must-gather-c8s97/must-gather-88wmh" Dec 15 13:34:21 crc kubenswrapper[4719]: I1215 13:34:21.934482 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/must-gather-88wmh" Dec 15 13:34:22 crc kubenswrapper[4719]: I1215 13:34:22.386688 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-c8s97/must-gather-88wmh"] Dec 15 13:34:22 crc kubenswrapper[4719]: I1215 13:34:22.428663 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/must-gather-88wmh" event={"ID":"2397d720-99e1-4da0-93ff-aac39122fadf","Type":"ContainerStarted","Data":"125353d7a43332e826905df316c5f2e7e4e1e792fd577e233b06c9f9ec077cbb"} Dec 15 13:34:31 crc kubenswrapper[4719]: I1215 13:34:31.477632 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:34:31 crc kubenswrapper[4719]: E1215 13:34:31.478648 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:34:32 crc kubenswrapper[4719]: I1215 13:34:32.537536 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/must-gather-88wmh" event={"ID":"2397d720-99e1-4da0-93ff-aac39122fadf","Type":"ContainerStarted","Data":"a7c34f9fe63109ec0bdf8bf2b6af765b24d417a1f9b30bbf9d89b9249f3103d7"} Dec 15 13:34:32 crc kubenswrapper[4719]: I1215 13:34:32.538090 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/must-gather-88wmh" event={"ID":"2397d720-99e1-4da0-93ff-aac39122fadf","Type":"ContainerStarted","Data":"2a69ac7487a338ebca56a28f9b35c53d4a63471163932223e8deebab40507a08"} Dec 15 13:34:32 crc kubenswrapper[4719]: I1215 13:34:32.556940 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c8s97/must-gather-88wmh" podStartSLOduration=2.384807904 podStartE2EDuration="11.556922868s" podCreationTimestamp="2025-12-15 13:34:21 +0000 UTC" firstStartedPulling="2025-12-15 13:34:22.404707378 +0000 UTC m=+4623.347000408" lastFinishedPulling="2025-12-15 13:34:31.576822342 +0000 UTC m=+4632.519115372" observedRunningTime="2025-12-15 13:34:32.550487806 +0000 UTC m=+4633.492780846" watchObservedRunningTime="2025-12-15 13:34:32.556922868 +0000 UTC m=+4633.499215898" Dec 15 13:34:36 crc kubenswrapper[4719]: I1215 13:34:36.949594 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c8s97/crc-debug-cpg78"] Dec 15 13:34:36 crc kubenswrapper[4719]: I1215 13:34:36.952892 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-cpg78" Dec 15 13:34:37 crc kubenswrapper[4719]: I1215 13:34:37.097149 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2db57\" (UniqueName: \"kubernetes.io/projected/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-kube-api-access-2db57\") pod \"crc-debug-cpg78\" (UID: \"b97bd8dd-5e64-458a-ba74-2c2327a0ec24\") " pod="openshift-must-gather-c8s97/crc-debug-cpg78" Dec 15 13:34:37 crc kubenswrapper[4719]: I1215 13:34:37.097527 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-host\") pod \"crc-debug-cpg78\" (UID: \"b97bd8dd-5e64-458a-ba74-2c2327a0ec24\") " pod="openshift-must-gather-c8s97/crc-debug-cpg78" Dec 15 13:34:37 crc kubenswrapper[4719]: I1215 13:34:37.199896 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-host\") pod \"crc-debug-cpg78\" (UID: \"b97bd8dd-5e64-458a-ba74-2c2327a0ec24\") " pod="openshift-must-gather-c8s97/crc-debug-cpg78" Dec 15 13:34:37 crc kubenswrapper[4719]: I1215 13:34:37.200031 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-host\") pod \"crc-debug-cpg78\" (UID: \"b97bd8dd-5e64-458a-ba74-2c2327a0ec24\") " pod="openshift-must-gather-c8s97/crc-debug-cpg78" Dec 15 13:34:37 crc kubenswrapper[4719]: I1215 13:34:37.200479 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2db57\" (UniqueName: \"kubernetes.io/projected/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-kube-api-access-2db57\") pod \"crc-debug-cpg78\" (UID: \"b97bd8dd-5e64-458a-ba74-2c2327a0ec24\") " pod="openshift-must-gather-c8s97/crc-debug-cpg78" Dec 15 13:34:37 crc kubenswrapper[4719]: I1215 13:34:37.232300 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2db57\" (UniqueName: \"kubernetes.io/projected/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-kube-api-access-2db57\") pod \"crc-debug-cpg78\" (UID: \"b97bd8dd-5e64-458a-ba74-2c2327a0ec24\") " pod="openshift-must-gather-c8s97/crc-debug-cpg78" Dec 15 13:34:37 crc kubenswrapper[4719]: I1215 13:34:37.272060 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-cpg78" Dec 15 13:34:37 crc kubenswrapper[4719]: W1215 13:34:37.299101 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb97bd8dd_5e64_458a_ba74_2c2327a0ec24.slice/crio-12116be15e5ee47ab54710f02d068832162417085e92227d47c0bb7f1c1e4d5b WatchSource:0}: Error finding container 12116be15e5ee47ab54710f02d068832162417085e92227d47c0bb7f1c1e4d5b: Status 404 returned error can't find the container with id 12116be15e5ee47ab54710f02d068832162417085e92227d47c0bb7f1c1e4d5b Dec 15 13:34:37 crc kubenswrapper[4719]: I1215 13:34:37.593804 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/crc-debug-cpg78" event={"ID":"b97bd8dd-5e64-458a-ba74-2c2327a0ec24","Type":"ContainerStarted","Data":"12116be15e5ee47ab54710f02d068832162417085e92227d47c0bb7f1c1e4d5b"} Dec 15 13:34:46 crc kubenswrapper[4719]: I1215 13:34:46.476877 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:34:46 crc kubenswrapper[4719]: E1215 13:34:46.477663 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:34:51 crc kubenswrapper[4719]: I1215 13:34:51.898283 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/crc-debug-cpg78" event={"ID":"b97bd8dd-5e64-458a-ba74-2c2327a0ec24","Type":"ContainerStarted","Data":"2001c130490d9811c3fec4ce675a4bdf200ee83ddc695fb8475c936662d5d1bb"} Dec 15 13:34:51 crc kubenswrapper[4719]: I1215 13:34:51.917689 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c8s97/crc-debug-cpg78" podStartSLOduration=2.263244773 podStartE2EDuration="15.917658971s" podCreationTimestamp="2025-12-15 13:34:36 +0000 UTC" firstStartedPulling="2025-12-15 13:34:37.305470009 +0000 UTC m=+4638.247763039" lastFinishedPulling="2025-12-15 13:34:50.959884207 +0000 UTC m=+4651.902177237" observedRunningTime="2025-12-15 13:34:51.913700577 +0000 UTC m=+4652.855993607" watchObservedRunningTime="2025-12-15 13:34:51.917658971 +0000 UTC m=+4652.859952001" Dec 15 13:35:01 crc kubenswrapper[4719]: I1215 13:35:01.476075 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:35:01 crc kubenswrapper[4719]: E1215 13:35:01.476959 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:35:13 crc kubenswrapper[4719]: I1215 13:35:13.476158 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:35:13 crc kubenswrapper[4719]: E1215 13:35:13.476919 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:35:26 crc kubenswrapper[4719]: I1215 13:35:26.475786 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:35:27 crc kubenswrapper[4719]: I1215 13:35:27.233262 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"e5f68797b715b1014a038536c351a5cc8510f4641ca0aafa31da234a3cd21cac"} Dec 15 13:35:31 crc kubenswrapper[4719]: I1215 13:35:30.781047 4719 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="3b614221-314c-4033-80d8-20738c1b92ae" containerName="galera" probeResult="failure" output="command timed out" Dec 15 13:35:31 crc kubenswrapper[4719]: I1215 13:35:31.181738 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="3b614221-314c-4033-80d8-20738c1b92ae" containerName="galera" probeResult="failure" output="command timed out" Dec 15 13:35:46 crc kubenswrapper[4719]: I1215 13:35:46.397513 4719 generic.go:334] "Generic (PLEG): container finished" podID="b97bd8dd-5e64-458a-ba74-2c2327a0ec24" containerID="2001c130490d9811c3fec4ce675a4bdf200ee83ddc695fb8475c936662d5d1bb" exitCode=0 Dec 15 13:35:46 crc kubenswrapper[4719]: I1215 13:35:46.397705 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/crc-debug-cpg78" event={"ID":"b97bd8dd-5e64-458a-ba74-2c2327a0ec24","Type":"ContainerDied","Data":"2001c130490d9811c3fec4ce675a4bdf200ee83ddc695fb8475c936662d5d1bb"} Dec 15 13:35:47 crc kubenswrapper[4719]: I1215 13:35:47.503473 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-cpg78" Dec 15 13:35:47 crc kubenswrapper[4719]: I1215 13:35:47.538005 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c8s97/crc-debug-cpg78"] Dec 15 13:35:47 crc kubenswrapper[4719]: I1215 13:35:47.546150 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c8s97/crc-debug-cpg78"] Dec 15 13:35:47 crc kubenswrapper[4719]: I1215 13:35:47.691083 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-host\") pod \"b97bd8dd-5e64-458a-ba74-2c2327a0ec24\" (UID: \"b97bd8dd-5e64-458a-ba74-2c2327a0ec24\") " Dec 15 13:35:47 crc kubenswrapper[4719]: I1215 13:35:47.691352 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-host" (OuterVolumeSpecName: "host") pod "b97bd8dd-5e64-458a-ba74-2c2327a0ec24" (UID: "b97bd8dd-5e64-458a-ba74-2c2327a0ec24"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 13:35:47 crc kubenswrapper[4719]: I1215 13:35:47.691398 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2db57\" (UniqueName: \"kubernetes.io/projected/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-kube-api-access-2db57\") pod \"b97bd8dd-5e64-458a-ba74-2c2327a0ec24\" (UID: \"b97bd8dd-5e64-458a-ba74-2c2327a0ec24\") " Dec 15 13:35:47 crc kubenswrapper[4719]: I1215 13:35:47.691697 4719 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-host\") on node \"crc\" DevicePath \"\"" Dec 15 13:35:47 crc kubenswrapper[4719]: I1215 13:35:47.704096 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-kube-api-access-2db57" (OuterVolumeSpecName: "kube-api-access-2db57") pod "b97bd8dd-5e64-458a-ba74-2c2327a0ec24" (UID: "b97bd8dd-5e64-458a-ba74-2c2327a0ec24"). InnerVolumeSpecName "kube-api-access-2db57". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:35:47 crc kubenswrapper[4719]: I1215 13:35:47.793842 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2db57\" (UniqueName: \"kubernetes.io/projected/b97bd8dd-5e64-458a-ba74-2c2327a0ec24-kube-api-access-2db57\") on node \"crc\" DevicePath \"\"" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.413373 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12116be15e5ee47ab54710f02d068832162417085e92227d47c0bb7f1c1e4d5b" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.413441 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-cpg78" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.758298 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c8s97/crc-debug-wqm84"] Dec 15 13:35:48 crc kubenswrapper[4719]: E1215 13:35:48.759111 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b97bd8dd-5e64-458a-ba74-2c2327a0ec24" containerName="container-00" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.759125 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b97bd8dd-5e64-458a-ba74-2c2327a0ec24" containerName="container-00" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.759290 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="b97bd8dd-5e64-458a-ba74-2c2327a0ec24" containerName="container-00" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.760124 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-wqm84" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.809575 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/860a1268-3991-4caa-b510-abe0cbe87968-host\") pod \"crc-debug-wqm84\" (UID: \"860a1268-3991-4caa-b510-abe0cbe87968\") " pod="openshift-must-gather-c8s97/crc-debug-wqm84" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.809658 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6mpw\" (UniqueName: \"kubernetes.io/projected/860a1268-3991-4caa-b510-abe0cbe87968-kube-api-access-l6mpw\") pod \"crc-debug-wqm84\" (UID: \"860a1268-3991-4caa-b510-abe0cbe87968\") " pod="openshift-must-gather-c8s97/crc-debug-wqm84" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.911519 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6mpw\" (UniqueName: \"kubernetes.io/projected/860a1268-3991-4caa-b510-abe0cbe87968-kube-api-access-l6mpw\") pod \"crc-debug-wqm84\" (UID: \"860a1268-3991-4caa-b510-abe0cbe87968\") " pod="openshift-must-gather-c8s97/crc-debug-wqm84" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.911722 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/860a1268-3991-4caa-b510-abe0cbe87968-host\") pod \"crc-debug-wqm84\" (UID: \"860a1268-3991-4caa-b510-abe0cbe87968\") " pod="openshift-must-gather-c8s97/crc-debug-wqm84" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.911825 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/860a1268-3991-4caa-b510-abe0cbe87968-host\") pod \"crc-debug-wqm84\" (UID: \"860a1268-3991-4caa-b510-abe0cbe87968\") " pod="openshift-must-gather-c8s97/crc-debug-wqm84" Dec 15 13:35:48 crc kubenswrapper[4719]: I1215 13:35:48.937712 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6mpw\" (UniqueName: \"kubernetes.io/projected/860a1268-3991-4caa-b510-abe0cbe87968-kube-api-access-l6mpw\") pod \"crc-debug-wqm84\" (UID: \"860a1268-3991-4caa-b510-abe0cbe87968\") " pod="openshift-must-gather-c8s97/crc-debug-wqm84" Dec 15 13:35:49 crc kubenswrapper[4719]: I1215 13:35:49.078247 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-wqm84" Dec 15 13:35:49 crc kubenswrapper[4719]: W1215 13:35:49.106657 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod860a1268_3991_4caa_b510_abe0cbe87968.slice/crio-0e0628e87abbaee13fce5505d02ea09dbe34730e9d770884b63ce67fb861f870 WatchSource:0}: Error finding container 0e0628e87abbaee13fce5505d02ea09dbe34730e9d770884b63ce67fb861f870: Status 404 returned error can't find the container with id 0e0628e87abbaee13fce5505d02ea09dbe34730e9d770884b63ce67fb861f870 Dec 15 13:35:49 crc kubenswrapper[4719]: I1215 13:35:49.423561 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/crc-debug-wqm84" event={"ID":"860a1268-3991-4caa-b510-abe0cbe87968","Type":"ContainerStarted","Data":"92b61a7e2674c517207964a2c542c7e81cd9a155997937f0d41cd8c5c4ae5e81"} Dec 15 13:35:49 crc kubenswrapper[4719]: I1215 13:35:49.423954 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/crc-debug-wqm84" event={"ID":"860a1268-3991-4caa-b510-abe0cbe87968","Type":"ContainerStarted","Data":"0e0628e87abbaee13fce5505d02ea09dbe34730e9d770884b63ce67fb861f870"} Dec 15 13:35:49 crc kubenswrapper[4719]: I1215 13:35:49.444931 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c8s97/crc-debug-wqm84" podStartSLOduration=1.444909669 podStartE2EDuration="1.444909669s" podCreationTimestamp="2025-12-15 13:35:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 13:35:49.437959231 +0000 UTC m=+4710.380252261" watchObservedRunningTime="2025-12-15 13:35:49.444909669 +0000 UTC m=+4710.387202709" Dec 15 13:35:49 crc kubenswrapper[4719]: I1215 13:35:49.489734 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b97bd8dd-5e64-458a-ba74-2c2327a0ec24" path="/var/lib/kubelet/pods/b97bd8dd-5e64-458a-ba74-2c2327a0ec24/volumes" Dec 15 13:35:50 crc kubenswrapper[4719]: I1215 13:35:50.434675 4719 generic.go:334] "Generic (PLEG): container finished" podID="860a1268-3991-4caa-b510-abe0cbe87968" containerID="92b61a7e2674c517207964a2c542c7e81cd9a155997937f0d41cd8c5c4ae5e81" exitCode=0 Dec 15 13:35:50 crc kubenswrapper[4719]: I1215 13:35:50.434765 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/crc-debug-wqm84" event={"ID":"860a1268-3991-4caa-b510-abe0cbe87968","Type":"ContainerDied","Data":"92b61a7e2674c517207964a2c542c7e81cd9a155997937f0d41cd8c5c4ae5e81"} Dec 15 13:35:51 crc kubenswrapper[4719]: I1215 13:35:51.546074 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-wqm84" Dec 15 13:35:51 crc kubenswrapper[4719]: I1215 13:35:51.657843 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/860a1268-3991-4caa-b510-abe0cbe87968-host\") pod \"860a1268-3991-4caa-b510-abe0cbe87968\" (UID: \"860a1268-3991-4caa-b510-abe0cbe87968\") " Dec 15 13:35:51 crc kubenswrapper[4719]: I1215 13:35:51.657950 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6mpw\" (UniqueName: \"kubernetes.io/projected/860a1268-3991-4caa-b510-abe0cbe87968-kube-api-access-l6mpw\") pod \"860a1268-3991-4caa-b510-abe0cbe87968\" (UID: \"860a1268-3991-4caa-b510-abe0cbe87968\") " Dec 15 13:35:51 crc kubenswrapper[4719]: I1215 13:35:51.658377 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/860a1268-3991-4caa-b510-abe0cbe87968-host" (OuterVolumeSpecName: "host") pod "860a1268-3991-4caa-b510-abe0cbe87968" (UID: "860a1268-3991-4caa-b510-abe0cbe87968"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 13:35:51 crc kubenswrapper[4719]: I1215 13:35:51.658687 4719 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/860a1268-3991-4caa-b510-abe0cbe87968-host\") on node \"crc\" DevicePath \"\"" Dec 15 13:35:51 crc kubenswrapper[4719]: I1215 13:35:51.667634 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/860a1268-3991-4caa-b510-abe0cbe87968-kube-api-access-l6mpw" (OuterVolumeSpecName: "kube-api-access-l6mpw") pod "860a1268-3991-4caa-b510-abe0cbe87968" (UID: "860a1268-3991-4caa-b510-abe0cbe87968"). InnerVolumeSpecName "kube-api-access-l6mpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:35:51 crc kubenswrapper[4719]: I1215 13:35:51.701407 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c8s97/crc-debug-wqm84"] Dec 15 13:35:51 crc kubenswrapper[4719]: I1215 13:35:51.710418 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c8s97/crc-debug-wqm84"] Dec 15 13:35:51 crc kubenswrapper[4719]: I1215 13:35:51.760184 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6mpw\" (UniqueName: \"kubernetes.io/projected/860a1268-3991-4caa-b510-abe0cbe87968-kube-api-access-l6mpw\") on node \"crc\" DevicePath \"\"" Dec 15 13:35:52 crc kubenswrapper[4719]: I1215 13:35:52.453644 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e0628e87abbaee13fce5505d02ea09dbe34730e9d770884b63ce67fb861f870" Dec 15 13:35:52 crc kubenswrapper[4719]: I1215 13:35:52.453682 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-wqm84" Dec 15 13:35:52 crc kubenswrapper[4719]: I1215 13:35:52.877592 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c8s97/crc-debug-b4fm2"] Dec 15 13:35:52 crc kubenswrapper[4719]: E1215 13:35:52.878319 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="860a1268-3991-4caa-b510-abe0cbe87968" containerName="container-00" Dec 15 13:35:52 crc kubenswrapper[4719]: I1215 13:35:52.878332 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="860a1268-3991-4caa-b510-abe0cbe87968" containerName="container-00" Dec 15 13:35:52 crc kubenswrapper[4719]: I1215 13:35:52.878496 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="860a1268-3991-4caa-b510-abe0cbe87968" containerName="container-00" Dec 15 13:35:52 crc kubenswrapper[4719]: I1215 13:35:52.879055 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-b4fm2" Dec 15 13:35:52 crc kubenswrapper[4719]: I1215 13:35:52.982455 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-host\") pod \"crc-debug-b4fm2\" (UID: \"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2\") " pod="openshift-must-gather-c8s97/crc-debug-b4fm2" Dec 15 13:35:52 crc kubenswrapper[4719]: I1215 13:35:52.982540 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bbqz\" (UniqueName: \"kubernetes.io/projected/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-kube-api-access-2bbqz\") pod \"crc-debug-b4fm2\" (UID: \"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2\") " pod="openshift-must-gather-c8s97/crc-debug-b4fm2" Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.084389 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-host\") pod \"crc-debug-b4fm2\" (UID: \"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2\") " pod="openshift-must-gather-c8s97/crc-debug-b4fm2" Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.084450 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bbqz\" (UniqueName: \"kubernetes.io/projected/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-kube-api-access-2bbqz\") pod \"crc-debug-b4fm2\" (UID: \"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2\") " pod="openshift-must-gather-c8s97/crc-debug-b4fm2" Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.084532 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-host\") pod \"crc-debug-b4fm2\" (UID: \"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2\") " pod="openshift-must-gather-c8s97/crc-debug-b4fm2" Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.101147 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bbqz\" (UniqueName: \"kubernetes.io/projected/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-kube-api-access-2bbqz\") pod \"crc-debug-b4fm2\" (UID: \"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2\") " pod="openshift-must-gather-c8s97/crc-debug-b4fm2" Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.194123 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-b4fm2" Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.461877 4719 generic.go:334] "Generic (PLEG): container finished" podID="eb3d78f1-3635-4bf4-b950-7ca6a2819bd2" containerID="d5544aaea30df47c9821427979bed7f25a88c98ecde6da87a001286a6fd2d1cf" exitCode=0 Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.462303 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/crc-debug-b4fm2" event={"ID":"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2","Type":"ContainerDied","Data":"d5544aaea30df47c9821427979bed7f25a88c98ecde6da87a001286a6fd2d1cf"} Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.462914 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/crc-debug-b4fm2" event={"ID":"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2","Type":"ContainerStarted","Data":"8196dbf133e55ad49fefa43bfdd778f106a48c22117a18eebcdeb7e774a4c42b"} Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.492499 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="860a1268-3991-4caa-b510-abe0cbe87968" path="/var/lib/kubelet/pods/860a1268-3991-4caa-b510-abe0cbe87968/volumes" Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.511399 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c8s97/crc-debug-b4fm2"] Dec 15 13:35:53 crc kubenswrapper[4719]: I1215 13:35:53.522295 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c8s97/crc-debug-b4fm2"] Dec 15 13:35:53 crc kubenswrapper[4719]: E1215 13:35:53.638533 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb3d78f1_3635_4bf4_b950_7ca6a2819bd2.slice/crio-d5544aaea30df47c9821427979bed7f25a88c98ecde6da87a001286a6fd2d1cf.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb3d78f1_3635_4bf4_b950_7ca6a2819bd2.slice/crio-conmon-d5544aaea30df47c9821427979bed7f25a88c98ecde6da87a001286a6fd2d1cf.scope\": RecentStats: unable to find data in memory cache]" Dec 15 13:35:54 crc kubenswrapper[4719]: I1215 13:35:54.572266 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-b4fm2" Dec 15 13:35:54 crc kubenswrapper[4719]: I1215 13:35:54.710111 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-host\") pod \"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2\" (UID: \"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2\") " Dec 15 13:35:54 crc kubenswrapper[4719]: I1215 13:35:54.710293 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bbqz\" (UniqueName: \"kubernetes.io/projected/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-kube-api-access-2bbqz\") pod \"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2\" (UID: \"eb3d78f1-3635-4bf4-b950-7ca6a2819bd2\") " Dec 15 13:35:54 crc kubenswrapper[4719]: I1215 13:35:54.711869 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-host" (OuterVolumeSpecName: "host") pod "eb3d78f1-3635-4bf4-b950-7ca6a2819bd2" (UID: "eb3d78f1-3635-4bf4-b950-7ca6a2819bd2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 13:35:54 crc kubenswrapper[4719]: I1215 13:35:54.725098 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-kube-api-access-2bbqz" (OuterVolumeSpecName: "kube-api-access-2bbqz") pod "eb3d78f1-3635-4bf4-b950-7ca6a2819bd2" (UID: "eb3d78f1-3635-4bf4-b950-7ca6a2819bd2"). InnerVolumeSpecName "kube-api-access-2bbqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:35:54 crc kubenswrapper[4719]: I1215 13:35:54.812821 4719 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-host\") on node \"crc\" DevicePath \"\"" Dec 15 13:35:54 crc kubenswrapper[4719]: I1215 13:35:54.813087 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bbqz\" (UniqueName: \"kubernetes.io/projected/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2-kube-api-access-2bbqz\") on node \"crc\" DevicePath \"\"" Dec 15 13:35:55 crc kubenswrapper[4719]: I1215 13:35:55.479940 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/crc-debug-b4fm2" Dec 15 13:35:55 crc kubenswrapper[4719]: I1215 13:35:55.489698 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb3d78f1-3635-4bf4-b950-7ca6a2819bd2" path="/var/lib/kubelet/pods/eb3d78f1-3635-4bf4-b950-7ca6a2819bd2/volumes" Dec 15 13:35:55 crc kubenswrapper[4719]: I1215 13:35:55.490480 4719 scope.go:117] "RemoveContainer" containerID="d5544aaea30df47c9821427979bed7f25a88c98ecde6da87a001286a6fd2d1cf" Dec 15 13:36:11 crc kubenswrapper[4719]: I1215 13:36:11.708321 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7ff7556d66-bpsdl_a996feea-1c4e-496f-a91b-2f07b852d22e/barbican-api/0.log" Dec 15 13:36:11 crc kubenswrapper[4719]: I1215 13:36:11.948006 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7ff7556d66-bpsdl_a996feea-1c4e-496f-a91b-2f07b852d22e/barbican-api-log/0.log" Dec 15 13:36:11 crc kubenswrapper[4719]: I1215 13:36:11.984908 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b87454b66-vfv5x_2da95a03-4fe4-4674-b56c-f697d4ab3045/barbican-keystone-listener/0.log" Dec 15 13:36:12 crc kubenswrapper[4719]: I1215 13:36:12.127592 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b87454b66-vfv5x_2da95a03-4fe4-4674-b56c-f697d4ab3045/barbican-keystone-listener-log/0.log" Dec 15 13:36:12 crc kubenswrapper[4719]: I1215 13:36:12.296767 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-769f77d87-tfk84_d2f1a60c-ffe6-4563-9466-2a3129be55ee/barbican-worker-log/0.log" Dec 15 13:36:12 crc kubenswrapper[4719]: I1215 13:36:12.317345 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-769f77d87-tfk84_d2f1a60c-ffe6-4563-9466-2a3129be55ee/barbican-worker/0.log" Dec 15 13:36:12 crc kubenswrapper[4719]: I1215 13:36:12.906399 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w_19e3b1bb-a052-47fd-a833-bed95ef223fc/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:12 crc kubenswrapper[4719]: I1215 13:36:12.983740 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0/ceilometer-central-agent/0.log" Dec 15 13:36:13 crc kubenswrapper[4719]: I1215 13:36:13.028481 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0/ceilometer-notification-agent/0.log" Dec 15 13:36:13 crc kubenswrapper[4719]: I1215 13:36:13.213362 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0/proxy-httpd/0.log" Dec 15 13:36:13 crc kubenswrapper[4719]: I1215 13:36:13.295297 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0/sg-core/0.log" Dec 15 13:36:13 crc kubenswrapper[4719]: I1215 13:36:13.326359 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6133c925-0cd7-489d-8b32-097e9b78ad70/cinder-api/0.log" Dec 15 13:36:13 crc kubenswrapper[4719]: I1215 13:36:13.511481 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6133c925-0cd7-489d-8b32-097e9b78ad70/cinder-api-log/0.log" Dec 15 13:36:13 crc kubenswrapper[4719]: I1215 13:36:13.624446 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d60e91f3-5bb1-4958-8489-e047f9a31c87/cinder-scheduler/0.log" Dec 15 13:36:13 crc kubenswrapper[4719]: I1215 13:36:13.651547 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d60e91f3-5bb1-4958-8489-e047f9a31c87/probe/0.log" Dec 15 13:36:13 crc kubenswrapper[4719]: I1215 13:36:13.826410 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-7q45z_2b3c8271-0d1c-4da0-b282-824968e884bb/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:13 crc kubenswrapper[4719]: I1215 13:36:13.956437 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj_611b65c2-c554-4f37-a644-41bcbd27ad46/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:14 crc kubenswrapper[4719]: I1215 13:36:14.133059 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-cvlz2_3cd29959-bbda-441f-901a-726fc4ebb4c4/init/0.log" Dec 15 13:36:14 crc kubenswrapper[4719]: I1215 13:36:14.462896 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-cvlz2_3cd29959-bbda-441f-901a-726fc4ebb4c4/init/0.log" Dec 15 13:36:14 crc kubenswrapper[4719]: I1215 13:36:14.539050 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct_b72d7eb2-0c87-4377-8338-118b53d43241/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:14 crc kubenswrapper[4719]: I1215 13:36:14.579549 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-cvlz2_3cd29959-bbda-441f-901a-726fc4ebb4c4/dnsmasq-dns/0.log" Dec 15 13:36:14 crc kubenswrapper[4719]: I1215 13:36:14.806768 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fc2e0fce-d6e0-47e6-bd88-34c45829436a/glance-httpd/0.log" Dec 15 13:36:14 crc kubenswrapper[4719]: I1215 13:36:14.847494 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fc2e0fce-d6e0-47e6-bd88-34c45829436a/glance-log/0.log" Dec 15 13:36:15 crc kubenswrapper[4719]: I1215 13:36:15.065284 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f26a2026-ce1a-4411-8e19-e1981be6f927/glance-httpd/0.log" Dec 15 13:36:15 crc kubenswrapper[4719]: I1215 13:36:15.109138 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f26a2026-ce1a-4411-8e19-e1981be6f927/glance-log/0.log" Dec 15 13:36:15 crc kubenswrapper[4719]: I1215 13:36:15.192425 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6bc5fb48-jqf6b_f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926/horizon/2.log" Dec 15 13:36:15 crc kubenswrapper[4719]: I1215 13:36:15.407074 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6bc5fb48-jqf6b_f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926/horizon/1.log" Dec 15 13:36:15 crc kubenswrapper[4719]: I1215 13:36:15.483279 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s_71b47db7-3079-4990-8a37-b02ea6f0368b/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:15 crc kubenswrapper[4719]: I1215 13:36:15.811914 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6bc5fb48-jqf6b_f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926/horizon-log/0.log" Dec 15 13:36:15 crc kubenswrapper[4719]: I1215 13:36:15.850545 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-gv6zg_2e7c3312-b561-431c-8794-c63098dbe33d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:16 crc kubenswrapper[4719]: I1215 13:36:16.300251 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29430061-9mfp2_32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992/keystone-cron/0.log" Dec 15 13:36:16 crc kubenswrapper[4719]: I1215 13:36:16.759528 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5c7789bfff-vvvqf_e087fc68-a428-4d63-b1c1-9d9a6b7b0c93/keystone-api/0.log" Dec 15 13:36:16 crc kubenswrapper[4719]: I1215 13:36:16.979973 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_9a236191-195f-4941-b740-ac7ee096858a/kube-state-metrics/0.log" Dec 15 13:36:17 crc kubenswrapper[4719]: I1215 13:36:17.008292 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql_c390d0c5-24b7-4810-baf1-d7b9fd9dd13f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:17 crc kubenswrapper[4719]: I1215 13:36:17.764141 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg_39093fe3-9551-4d58-9cd0-607f1694a71c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:18 crc kubenswrapper[4719]: I1215 13:36:18.170534 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5668b96c97-48pkv_4ddfe7f4-2192-4520-bfae-028a2c41e459/neutron-httpd/0.log" Dec 15 13:36:18 crc kubenswrapper[4719]: I1215 13:36:18.422638 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5668b96c97-48pkv_4ddfe7f4-2192-4520-bfae-028a2c41e459/neutron-api/0.log" Dec 15 13:36:19 crc kubenswrapper[4719]: I1215 13:36:19.053493 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d2da200b-26c0-4242-a443-fa579e6da8da/nova-cell0-conductor-conductor/0.log" Dec 15 13:36:19 crc kubenswrapper[4719]: I1215 13:36:19.178529 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_36c7f0b4-980a-4543-a7bc-38e6115a9d53/nova-cell1-conductor-conductor/0.log" Dec 15 13:36:19 crc kubenswrapper[4719]: I1215 13:36:19.552547 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_59d59387-003d-47ae-a322-48787db59367/nova-api-log/0.log" Dec 15 13:36:19 crc kubenswrapper[4719]: I1215 13:36:19.645572 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ccc69b5f-a98b-4188-b08f-4cb6f58c085c/nova-cell1-novncproxy-novncproxy/0.log" Dec 15 13:36:19 crc kubenswrapper[4719]: I1215 13:36:19.912447 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-hvqz9_48c18077-569a-4bbd-ace2-64dbf06bcbee/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:19 crc kubenswrapper[4719]: I1215 13:36:19.940691 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_59d59387-003d-47ae-a322-48787db59367/nova-api-api/0.log" Dec 15 13:36:20 crc kubenswrapper[4719]: I1215 13:36:20.125895 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_45c8d521-b7f2-4f93-a6e8-94d6aaf85390/nova-metadata-log/0.log" Dec 15 13:36:20 crc kubenswrapper[4719]: I1215 13:36:20.405974 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9655f31a-6eae-461b-932a-0b1b4ca56048/mysql-bootstrap/0.log" Dec 15 13:36:20 crc kubenswrapper[4719]: I1215 13:36:20.765596 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_0945867e-d6c4-4f3e-976d-b2ae68497a11/nova-scheduler-scheduler/0.log" Dec 15 13:36:21 crc kubenswrapper[4719]: I1215 13:36:21.012344 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9655f31a-6eae-461b-932a-0b1b4ca56048/mysql-bootstrap/0.log" Dec 15 13:36:21 crc kubenswrapper[4719]: I1215 13:36:21.056846 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9655f31a-6eae-461b-932a-0b1b4ca56048/galera/0.log" Dec 15 13:36:21 crc kubenswrapper[4719]: I1215 13:36:21.338785 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b614221-314c-4033-80d8-20738c1b92ae/mysql-bootstrap/0.log" Dec 15 13:36:21 crc kubenswrapper[4719]: I1215 13:36:21.506776 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b614221-314c-4033-80d8-20738c1b92ae/mysql-bootstrap/0.log" Dec 15 13:36:21 crc kubenswrapper[4719]: I1215 13:36:21.546795 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b614221-314c-4033-80d8-20738c1b92ae/galera/0.log" Dec 15 13:36:21 crc kubenswrapper[4719]: I1215 13:36:21.864741 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0b238728-d6aa-4b27-b48b-beb0cc7ff437/openstackclient/0.log" Dec 15 13:36:21 crc kubenswrapper[4719]: I1215 13:36:21.896040 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1/memcached/0.log" Dec 15 13:36:21 crc kubenswrapper[4719]: I1215 13:36:21.997998 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_45c8d521-b7f2-4f93-a6e8-94d6aaf85390/nova-metadata-metadata/0.log" Dec 15 13:36:22 crc kubenswrapper[4719]: I1215 13:36:22.005366 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4ht64_4c7986d1-97c7-4b8c-8b2f-594fafc24991/ovn-controller/0.log" Dec 15 13:36:22 crc kubenswrapper[4719]: I1215 13:36:22.102888 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-84b56_8dc6d84f-343b-456a-bad8-3171dccc2b9d/openstack-network-exporter/0.log" Dec 15 13:36:22 crc kubenswrapper[4719]: I1215 13:36:22.241279 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hrrfr_d60bc22b-09f1-4b84-84ad-a9ee31a18a43/ovsdb-server-init/0.log" Dec 15 13:36:22 crc kubenswrapper[4719]: I1215 13:36:22.447009 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hrrfr_d60bc22b-09f1-4b84-84ad-a9ee31a18a43/ovsdb-server-init/0.log" Dec 15 13:36:22 crc kubenswrapper[4719]: I1215 13:36:22.490565 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hrrfr_d60bc22b-09f1-4b84-84ad-a9ee31a18a43/ovsdb-server/0.log" Dec 15 13:36:22 crc kubenswrapper[4719]: I1215 13:36:22.493176 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hrrfr_d60bc22b-09f1-4b84-84ad-a9ee31a18a43/ovs-vswitchd/0.log" Dec 15 13:36:22 crc kubenswrapper[4719]: I1215 13:36:22.719186 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-nrfr7_87e0fa54-5d1a-4e53-aef8-257af1ebd4d3/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:22 crc kubenswrapper[4719]: I1215 13:36:22.729506 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a0b079a1-3e14-4457-aac6-213038651316/ovn-northd/0.log" Dec 15 13:36:22 crc kubenswrapper[4719]: I1215 13:36:22.790331 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a0b079a1-3e14-4457-aac6-213038651316/openstack-network-exporter/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.040837 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_66e0622c-a031-4927-83f3-d7d4c8dc0233/ovsdbserver-nb/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.053415 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_66e0622c-a031-4927-83f3-d7d4c8dc0233/openstack-network-exporter/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.116066 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_66f93b1d-4c36-4f90-ac6c-28b302a8c34b/openstack-network-exporter/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.241335 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_66f93b1d-4c36-4f90-ac6c-28b302a8c34b/ovsdbserver-sb/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.430552 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9d9f8bcf4-hj8fd_ac288046-bbc8-4521-84b8-c16e7d096217/placement-api/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.549457 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9d9f8bcf4-hj8fd_ac288046-bbc8-4521-84b8-c16e7d096217/placement-log/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.550932 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82c8c133-e1d6-4b56-b745-1aed017d0bfb/setup-container/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.711364 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82c8c133-e1d6-4b56-b745-1aed017d0bfb/setup-container/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.738446 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82c8c133-e1d6-4b56-b745-1aed017d0bfb/rabbitmq/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.800775 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d4163ed6-9550-4d10-baab-b92859fa07df/setup-container/0.log" Dec 15 13:36:23 crc kubenswrapper[4719]: I1215 13:36:23.998082 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d4163ed6-9550-4d10-baab-b92859fa07df/setup-container/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.058729 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk_f1474b5e-e320-4f22-97ae-729dbdd97939/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.148638 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d4163ed6-9550-4d10-baab-b92859fa07df/rabbitmq/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.224277 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-xjgnt_ed08a59f-ed8b-4195-b916-0d04c3e01f9f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.307882 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s_588db7b9-6a66-4c02-be83-8766e47c211c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.412738 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-kfrcr_57d95048-e4af-466c-84ff-f2d00d3b5f58/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.649468 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-xhxtk_c7b477ac-26a4-4ddb-af83-57d931fe0e5a/ssh-known-hosts-edpm-deployment/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.757460 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7f6dc9444c-2jcmv_60352f3a-8ee6-498d-9ee3-152cc4ab562f/proxy-httpd/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.760506 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7f6dc9444c-2jcmv_60352f3a-8ee6-498d-9ee3-152cc4ab562f/proxy-server/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.868476 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-dcbfk_fb01ec8f-cc85-4146-9677-344c14c2a545/swift-ring-rebalance/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.953891 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/account-auditor/0.log" Dec 15 13:36:24 crc kubenswrapper[4719]: I1215 13:36:24.982382 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/account-reaper/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.042538 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/account-replicator/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.127950 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/account-server/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.189588 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/container-replicator/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.209342 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/container-auditor/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.211848 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/container-server/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.268927 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/container-updater/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.379609 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/object-expirer/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.420476 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/object-server/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.421145 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/object-auditor/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.464671 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/object-replicator/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.523202 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/object-updater/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.589181 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/rsync/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.663778 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/swift-recon-cron/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.819388 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb_e70ef11b-7e5a-47c9-9f63-15dd69326073/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:25 crc kubenswrapper[4719]: I1215 13:36:25.877018 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_8b81902c-c979-4e03-a40f-cac54a63d6db/tempest-tests-tempest-tests-runner/0.log" Dec 15 13:36:26 crc kubenswrapper[4719]: I1215 13:36:26.004726 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_d347fbd6-e218-4153-be3f-f912f19123bb/test-operator-logs-container/0.log" Dec 15 13:36:26 crc kubenswrapper[4719]: I1215 13:36:26.253606 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl_2c709af4-4a8c-4cb9-955e-37dfe843569b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:36:53 crc kubenswrapper[4719]: I1215 13:36:53.041011 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/util/0.log" Dec 15 13:36:53 crc kubenswrapper[4719]: I1215 13:36:53.292223 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/pull/0.log" Dec 15 13:36:53 crc kubenswrapper[4719]: I1215 13:36:53.334678 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/pull/0.log" Dec 15 13:36:53 crc kubenswrapper[4719]: I1215 13:36:53.356349 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/util/0.log" Dec 15 13:36:53 crc kubenswrapper[4719]: I1215 13:36:53.496495 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/util/0.log" Dec 15 13:36:53 crc kubenswrapper[4719]: I1215 13:36:53.547061 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/pull/0.log" Dec 15 13:36:53 crc kubenswrapper[4719]: I1215 13:36:53.562500 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/extract/0.log" Dec 15 13:36:53 crc kubenswrapper[4719]: I1215 13:36:53.824814 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-95949466-k2fz4_5c7fa683-b59c-45ef-bc29-92400f43e4b8/manager/0.log" Dec 15 13:36:53 crc kubenswrapper[4719]: I1215 13:36:53.901828 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5cf45c46bd-pmmwk_7f47c007-e5bf-47dd-8905-69ed7cfc3a9a/manager/0.log" Dec 15 13:36:54 crc kubenswrapper[4719]: I1215 13:36:54.116215 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66f8b87655-5td8v_7ced32f3-ca3c-406c-966b-93be87cd6d25/manager/0.log" Dec 15 13:36:54 crc kubenswrapper[4719]: I1215 13:36:54.243707 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-767f9d7567-2866h_345b77d5-0419-433e-b7d4-053a0b6fcf86/manager/0.log" Dec 15 13:36:54 crc kubenswrapper[4719]: I1215 13:36:54.401393 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-59b8dcb766-llbls_6cdd3328-cd4d-4e90-8ab9-d0953cab73de/manager/0.log" Dec 15 13:36:54 crc kubenswrapper[4719]: I1215 13:36:54.558276 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6ccf486b9-scntm_15643ba0-8573-4f51-8ee6-9fd78e10b6a0/manager/0.log" Dec 15 13:36:54 crc kubenswrapper[4719]: I1215 13:36:54.871956 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-58944d7758-wh4dh_b6efe10d-9999-4340-822b-48fb8fb99f14/manager/0.log" Dec 15 13:36:55 crc kubenswrapper[4719]: I1215 13:36:55.000970 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-f458558d7-zkv6s_e6adc1e0-5c73-431d-a94d-e2eabc16dd03/manager/0.log" Dec 15 13:36:55 crc kubenswrapper[4719]: I1215 13:36:55.018401 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5c7cbf548f-5slz5_9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc/manager/0.log" Dec 15 13:36:55 crc kubenswrapper[4719]: I1215 13:36:55.224921 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5fdd9786f7-89zrh_571dd62a-58e8-4dab-ad04-a95621a65078/manager/0.log" Dec 15 13:36:55 crc kubenswrapper[4719]: I1215 13:36:55.283479 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f76f4954c-zhvw6_def1648e-ccf3-4a95-9a44-92bf0f83cb49/manager/0.log" Dec 15 13:36:55 crc kubenswrapper[4719]: I1215 13:36:55.487273 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7cd87b778f-hb66s_83a0cff1-8c43-4bb1-aa4f-66c4662fd235/manager/0.log" Dec 15 13:36:55 crc kubenswrapper[4719]: I1215 13:36:55.615068 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5fbbf8b6cc-c9szj_828d7a4d-0956-40fc-967f-e52d1701953c/manager/0.log" Dec 15 13:36:55 crc kubenswrapper[4719]: I1215 13:36:55.785234 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-68c649d9d-p7z6f_3b66851a-75d1-46ea-8d06-ad7385fdff6b/manager/0.log" Dec 15 13:36:55 crc kubenswrapper[4719]: I1215 13:36:55.902244 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556_24c5b173-c013-45e1-9563-1e49556b23a5/manager/0.log" Dec 15 13:36:56 crc kubenswrapper[4719]: I1215 13:36:56.388146 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zc4xm_bcaba66e-19b2-4c6f-9ec2-e545a7531624/registry-server/0.log" Dec 15 13:36:56 crc kubenswrapper[4719]: I1215 13:36:56.422516 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-66c467f856-n9b75_131429f3-e15e-4107-af0e-b2d4c01f87af/operator/0.log" Dec 15 13:36:56 crc kubenswrapper[4719]: I1215 13:36:56.945181 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8665b56d78-wslbg_9adcd81c-e72b-44b8-bd4f-36e49c8e5c82/manager/0.log" Dec 15 13:36:56 crc kubenswrapper[4719]: I1215 13:36:56.982724 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bf6d4f946-gn5n9_d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e/manager/0.log" Dec 15 13:36:57 crc kubenswrapper[4719]: I1215 13:36:57.236866 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-jzg2t_0c36396e-131f-4154-b82e-490b36d7fa63/operator/0.log" Dec 15 13:36:57 crc kubenswrapper[4719]: I1215 13:36:57.422374 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5c6df8f9-d9d67_c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157/manager/0.log" Dec 15 13:36:57 crc kubenswrapper[4719]: I1215 13:36:57.737436 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-97d456b9-qxhfl_375826ba-0785-4935-96c2-a2e5345aaf45/manager/0.log" Dec 15 13:36:57 crc kubenswrapper[4719]: I1215 13:36:57.763609 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7b4b554f64-x55hj_db19bab3-733c-4ef1-8f46-df4cf5842701/manager/0.log" Dec 15 13:36:57 crc kubenswrapper[4719]: I1215 13:36:57.958260 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-55f78b7c4c-292vp_a9ebdf36-2095-47b8-8edf-89036ffdee84/manager/0.log" Dec 15 13:36:57 crc kubenswrapper[4719]: I1215 13:36:57.959408 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-756ccf86c7-l98v4_72cd4bb7-5783-492b-bae4-1a9fa633cd7b/manager/0.log" Dec 15 13:37:18 crc kubenswrapper[4719]: I1215 13:37:18.068434 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ngv2q_d620f9fd-1374-4f93-972d-93f2b9173471/control-plane-machine-set-operator/0.log" Dec 15 13:37:18 crc kubenswrapper[4719]: I1215 13:37:18.238685 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c5df6_0dd6e306-1355-43dd-895f-598d3d873a42/kube-rbac-proxy/0.log" Dec 15 13:37:18 crc kubenswrapper[4719]: I1215 13:37:18.324108 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c5df6_0dd6e306-1355-43dd-895f-598d3d873a42/machine-api-operator/0.log" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.052593 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-6mllc_87e97929-9f4b-4e2e-bbc2-e9295954a8ae/cert-manager-controller/0.log" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.146058 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9jqf7"] Dec 15 13:37:32 crc kubenswrapper[4719]: E1215 13:37:32.146549 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb3d78f1-3635-4bf4-b950-7ca6a2819bd2" containerName="container-00" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.146572 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb3d78f1-3635-4bf4-b950-7ca6a2819bd2" containerName="container-00" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.146771 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb3d78f1-3635-4bf4-b950-7ca6a2819bd2" containerName="container-00" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.148366 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.154993 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jqf7"] Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.293420 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8tfl\" (UniqueName: \"kubernetes.io/projected/ddd690d7-223c-49c2-b743-79caec2762bb-kube-api-access-l8tfl\") pod \"redhat-marketplace-9jqf7\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.294025 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-utilities\") pod \"redhat-marketplace-9jqf7\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.294134 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-catalog-content\") pod \"redhat-marketplace-9jqf7\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.340536 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t69mj"] Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.346811 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.356541 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t69mj"] Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.392802 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-pjzlc_72cfd6e6-a544-4002-81ab-b57dfd2eaf94/cert-manager-cainjector/0.log" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.395640 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-catalog-content\") pod \"redhat-marketplace-9jqf7\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.395696 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8tfl\" (UniqueName: \"kubernetes.io/projected/ddd690d7-223c-49c2-b743-79caec2762bb-kube-api-access-l8tfl\") pod \"redhat-marketplace-9jqf7\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.395919 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-utilities\") pod \"redhat-marketplace-9jqf7\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.396195 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-catalog-content\") pod \"redhat-marketplace-9jqf7\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.412255 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-utilities\") pod \"redhat-marketplace-9jqf7\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.456332 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8tfl\" (UniqueName: \"kubernetes.io/projected/ddd690d7-223c-49c2-b743-79caec2762bb-kube-api-access-l8tfl\") pod \"redhat-marketplace-9jqf7\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.469361 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.497711 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-catalog-content\") pod \"certified-operators-t69mj\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.497785 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-utilities\") pod \"certified-operators-t69mj\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.497837 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mkxv\" (UniqueName: \"kubernetes.io/projected/01e39a53-c817-412f-bf0e-e5293c6b0a9f-kube-api-access-8mkxv\") pod \"certified-operators-t69mj\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.541441 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-nkj5k_994c9412-da12-497c-8849-a5948a36e975/cert-manager-webhook/0.log" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.599295 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-catalog-content\") pod \"certified-operators-t69mj\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.600062 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-utilities\") pod \"certified-operators-t69mj\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.600160 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mkxv\" (UniqueName: \"kubernetes.io/projected/01e39a53-c817-412f-bf0e-e5293c6b0a9f-kube-api-access-8mkxv\") pod \"certified-operators-t69mj\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.599898 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-catalog-content\") pod \"certified-operators-t69mj\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.602010 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-utilities\") pod \"certified-operators-t69mj\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.623114 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mkxv\" (UniqueName: \"kubernetes.io/projected/01e39a53-c817-412f-bf0e-e5293c6b0a9f-kube-api-access-8mkxv\") pod \"certified-operators-t69mj\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.712334 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:32 crc kubenswrapper[4719]: I1215 13:37:32.858828 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jqf7"] Dec 15 13:37:33 crc kubenswrapper[4719]: I1215 13:37:33.352876 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jqf7" event={"ID":"ddd690d7-223c-49c2-b743-79caec2762bb","Type":"ContainerStarted","Data":"556a14ab7eac595654ed7e301c4c6c5889c3a9629091e7b02568bb73a304f6f4"} Dec 15 13:37:33 crc kubenswrapper[4719]: I1215 13:37:33.731782 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t69mj"] Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.362110 4719 generic.go:334] "Generic (PLEG): container finished" podID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerID="36ee25b082e1b00b8348038a9d74c4531f97e48b2a8e5bb01d4ddc1d985960c7" exitCode=0 Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.362201 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t69mj" event={"ID":"01e39a53-c817-412f-bf0e-e5293c6b0a9f","Type":"ContainerDied","Data":"36ee25b082e1b00b8348038a9d74c4531f97e48b2a8e5bb01d4ddc1d985960c7"} Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.362486 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t69mj" event={"ID":"01e39a53-c817-412f-bf0e-e5293c6b0a9f","Type":"ContainerStarted","Data":"0e1f21555e3dcedc6747bc040157826bf0527eec685636acd692ed8eec7e2835"} Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.364388 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.365229 4719 generic.go:334] "Generic (PLEG): container finished" podID="ddd690d7-223c-49c2-b743-79caec2762bb" containerID="5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43" exitCode=0 Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.365264 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jqf7" event={"ID":"ddd690d7-223c-49c2-b743-79caec2762bb","Type":"ContainerDied","Data":"5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43"} Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.746948 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6sxjs"] Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.749472 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.757011 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6sxjs"] Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.855233 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-catalog-content\") pod \"community-operators-6sxjs\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.855634 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v59zn\" (UniqueName: \"kubernetes.io/projected/3e5797a5-9b7b-4613-8522-e223c3fe577a-kube-api-access-v59zn\") pod \"community-operators-6sxjs\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.855757 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-utilities\") pod \"community-operators-6sxjs\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.957467 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v59zn\" (UniqueName: \"kubernetes.io/projected/3e5797a5-9b7b-4613-8522-e223c3fe577a-kube-api-access-v59zn\") pod \"community-operators-6sxjs\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.957572 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-utilities\") pod \"community-operators-6sxjs\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.957648 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-catalog-content\") pod \"community-operators-6sxjs\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.958305 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-utilities\") pod \"community-operators-6sxjs\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.958339 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-catalog-content\") pod \"community-operators-6sxjs\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:34 crc kubenswrapper[4719]: I1215 13:37:34.982065 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v59zn\" (UniqueName: \"kubernetes.io/projected/3e5797a5-9b7b-4613-8522-e223c3fe577a-kube-api-access-v59zn\") pod \"community-operators-6sxjs\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:35 crc kubenswrapper[4719]: I1215 13:37:35.121200 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:35 crc kubenswrapper[4719]: I1215 13:37:35.410248 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jqf7" event={"ID":"ddd690d7-223c-49c2-b743-79caec2762bb","Type":"ContainerStarted","Data":"5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b"} Dec 15 13:37:35 crc kubenswrapper[4719]: I1215 13:37:35.785240 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6sxjs"] Dec 15 13:37:36 crc kubenswrapper[4719]: E1215 13:37:36.197390 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e5797a5_9b7b_4613_8522_e223c3fe577a.slice/crio-conmon-72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7.scope\": RecentStats: unable to find data in memory cache]" Dec 15 13:37:36 crc kubenswrapper[4719]: I1215 13:37:36.423442 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t69mj" event={"ID":"01e39a53-c817-412f-bf0e-e5293c6b0a9f","Type":"ContainerStarted","Data":"084f036166365287fa10de6ef747b24827f704b6f4cce3545f522183d5ba5c60"} Dec 15 13:37:36 crc kubenswrapper[4719]: I1215 13:37:36.427384 4719 generic.go:334] "Generic (PLEG): container finished" podID="ddd690d7-223c-49c2-b743-79caec2762bb" containerID="5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b" exitCode=0 Dec 15 13:37:36 crc kubenswrapper[4719]: I1215 13:37:36.427496 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jqf7" event={"ID":"ddd690d7-223c-49c2-b743-79caec2762bb","Type":"ContainerDied","Data":"5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b"} Dec 15 13:37:36 crc kubenswrapper[4719]: I1215 13:37:36.430141 4719 generic.go:334] "Generic (PLEG): container finished" podID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerID="72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7" exitCode=0 Dec 15 13:37:36 crc kubenswrapper[4719]: I1215 13:37:36.430196 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6sxjs" event={"ID":"3e5797a5-9b7b-4613-8522-e223c3fe577a","Type":"ContainerDied","Data":"72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7"} Dec 15 13:37:36 crc kubenswrapper[4719]: I1215 13:37:36.430232 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6sxjs" event={"ID":"3e5797a5-9b7b-4613-8522-e223c3fe577a","Type":"ContainerStarted","Data":"33db60fbd165302b3841fdf33bb88320aa1c9771557011cfd2a603c27530082e"} Dec 15 13:37:38 crc kubenswrapper[4719]: I1215 13:37:38.447734 4719 generic.go:334] "Generic (PLEG): container finished" podID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerID="084f036166365287fa10de6ef747b24827f704b6f4cce3545f522183d5ba5c60" exitCode=0 Dec 15 13:37:38 crc kubenswrapper[4719]: I1215 13:37:38.447806 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t69mj" event={"ID":"01e39a53-c817-412f-bf0e-e5293c6b0a9f","Type":"ContainerDied","Data":"084f036166365287fa10de6ef747b24827f704b6f4cce3545f522183d5ba5c60"} Dec 15 13:37:39 crc kubenswrapper[4719]: I1215 13:37:39.463766 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jqf7" event={"ID":"ddd690d7-223c-49c2-b743-79caec2762bb","Type":"ContainerStarted","Data":"d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425"} Dec 15 13:37:39 crc kubenswrapper[4719]: I1215 13:37:39.471054 4719 generic.go:334] "Generic (PLEG): container finished" podID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerID="b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44" exitCode=0 Dec 15 13:37:39 crc kubenswrapper[4719]: I1215 13:37:39.471099 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6sxjs" event={"ID":"3e5797a5-9b7b-4613-8522-e223c3fe577a","Type":"ContainerDied","Data":"b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44"} Dec 15 13:37:39 crc kubenswrapper[4719]: I1215 13:37:39.502068 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9jqf7" podStartSLOduration=3.609556488 podStartE2EDuration="7.50204539s" podCreationTimestamp="2025-12-15 13:37:32 +0000 UTC" firstStartedPulling="2025-12-15 13:37:34.367261471 +0000 UTC m=+4815.309554501" lastFinishedPulling="2025-12-15 13:37:38.259750373 +0000 UTC m=+4819.202043403" observedRunningTime="2025-12-15 13:37:39.490795286 +0000 UTC m=+4820.433088316" watchObservedRunningTime="2025-12-15 13:37:39.50204539 +0000 UTC m=+4820.444338420" Dec 15 13:37:40 crc kubenswrapper[4719]: I1215 13:37:40.482385 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6sxjs" event={"ID":"3e5797a5-9b7b-4613-8522-e223c3fe577a","Type":"ContainerStarted","Data":"16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe"} Dec 15 13:37:40 crc kubenswrapper[4719]: I1215 13:37:40.485430 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t69mj" event={"ID":"01e39a53-c817-412f-bf0e-e5293c6b0a9f","Type":"ContainerStarted","Data":"e2ab44da35b1f75d59fab94a314c1f17149fd244dd512d9e50f8ab7965121564"} Dec 15 13:37:40 crc kubenswrapper[4719]: I1215 13:37:40.510391 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6sxjs" podStartSLOduration=2.752849735 podStartE2EDuration="6.510372407s" podCreationTimestamp="2025-12-15 13:37:34 +0000 UTC" firstStartedPulling="2025-12-15 13:37:36.432464044 +0000 UTC m=+4817.374757074" lastFinishedPulling="2025-12-15 13:37:40.189986706 +0000 UTC m=+4821.132279746" observedRunningTime="2025-12-15 13:37:40.502015263 +0000 UTC m=+4821.444308293" watchObservedRunningTime="2025-12-15 13:37:40.510372407 +0000 UTC m=+4821.452665437" Dec 15 13:37:40 crc kubenswrapper[4719]: I1215 13:37:40.529022 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t69mj" podStartSLOduration=3.6734646 podStartE2EDuration="8.529004504s" podCreationTimestamp="2025-12-15 13:37:32 +0000 UTC" firstStartedPulling="2025-12-15 13:37:34.364200205 +0000 UTC m=+4815.306493235" lastFinishedPulling="2025-12-15 13:37:39.219740109 +0000 UTC m=+4820.162033139" observedRunningTime="2025-12-15 13:37:40.524346647 +0000 UTC m=+4821.466639697" watchObservedRunningTime="2025-12-15 13:37:40.529004504 +0000 UTC m=+4821.471297534" Dec 15 13:37:42 crc kubenswrapper[4719]: I1215 13:37:42.470700 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:42 crc kubenswrapper[4719]: I1215 13:37:42.471090 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:42 crc kubenswrapper[4719]: I1215 13:37:42.623114 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:42 crc kubenswrapper[4719]: I1215 13:37:42.713422 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:42 crc kubenswrapper[4719]: I1215 13:37:42.713475 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:42 crc kubenswrapper[4719]: I1215 13:37:42.768072 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:45 crc kubenswrapper[4719]: I1215 13:37:45.121475 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:45 crc kubenswrapper[4719]: I1215 13:37:45.121830 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:45 crc kubenswrapper[4719]: I1215 13:37:45.167733 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:45 crc kubenswrapper[4719]: I1215 13:37:45.594287 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:47 crc kubenswrapper[4719]: I1215 13:37:47.134150 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6sxjs"] Dec 15 13:37:47 crc kubenswrapper[4719]: I1215 13:37:47.574675 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6sxjs" podUID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerName="registry-server" containerID="cri-o://16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe" gracePeriod=2 Dec 15 13:37:47 crc kubenswrapper[4719]: I1215 13:37:47.665577 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6ff7998486-ljnvk_5b49fba8-f261-4a35-9aa0-669b7e7e70ab/nmstate-console-plugin/0.log" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.095452 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-trxk8_6524d6cd-421e-415d-826b-7ec166bc3915/nmstate-handler/0.log" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.122774 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-vwmdc_1fd5707c-b363-45cc-b2b3-168817ec1526/kube-rbac-proxy/0.log" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.122831 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.200651 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-vwmdc_1fd5707c-b363-45cc-b2b3-168817ec1526/nmstate-metrics/0.log" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.243441 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-catalog-content\") pod \"3e5797a5-9b7b-4613-8522-e223c3fe577a\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.243943 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-utilities\") pod \"3e5797a5-9b7b-4613-8522-e223c3fe577a\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.244790 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-utilities" (OuterVolumeSpecName: "utilities") pod "3e5797a5-9b7b-4613-8522-e223c3fe577a" (UID: "3e5797a5-9b7b-4613-8522-e223c3fe577a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.245657 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v59zn\" (UniqueName: \"kubernetes.io/projected/3e5797a5-9b7b-4613-8522-e223c3fe577a-kube-api-access-v59zn\") pod \"3e5797a5-9b7b-4613-8522-e223c3fe577a\" (UID: \"3e5797a5-9b7b-4613-8522-e223c3fe577a\") " Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.246196 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.256065 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e5797a5-9b7b-4613-8522-e223c3fe577a-kube-api-access-v59zn" (OuterVolumeSpecName: "kube-api-access-v59zn") pod "3e5797a5-9b7b-4613-8522-e223c3fe577a" (UID: "3e5797a5-9b7b-4613-8522-e223c3fe577a"). InnerVolumeSpecName "kube-api-access-v59zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.332813 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e5797a5-9b7b-4613-8522-e223c3fe577a" (UID: "3e5797a5-9b7b-4613-8522-e223c3fe577a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.348500 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5797a5-9b7b-4613-8522-e223c3fe577a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.348738 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v59zn\" (UniqueName: \"kubernetes.io/projected/3e5797a5-9b7b-4613-8522-e223c3fe577a-kube-api-access-v59zn\") on node \"crc\" DevicePath \"\"" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.451433 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-6769fb99d-h4ptq_91b08102-b558-4d78-ade6-0360f6e7846a/nmstate-operator/0.log" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.487012 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-f8fb84555-kqgm4_cee112f9-d46f-410b-bfc5-5d6cff881bb4/nmstate-webhook/0.log" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.594705 4719 generic.go:334] "Generic (PLEG): container finished" podID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerID="16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe" exitCode=0 Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.594752 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6sxjs" event={"ID":"3e5797a5-9b7b-4613-8522-e223c3fe577a","Type":"ContainerDied","Data":"16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe"} Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.594783 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6sxjs" event={"ID":"3e5797a5-9b7b-4613-8522-e223c3fe577a","Type":"ContainerDied","Data":"33db60fbd165302b3841fdf33bb88320aa1c9771557011cfd2a603c27530082e"} Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.594807 4719 scope.go:117] "RemoveContainer" containerID="16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.594980 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6sxjs" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.632101 4719 scope.go:117] "RemoveContainer" containerID="b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.641924 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6sxjs"] Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.653264 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6sxjs"] Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.687051 4719 scope.go:117] "RemoveContainer" containerID="72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.720567 4719 scope.go:117] "RemoveContainer" containerID="16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe" Dec 15 13:37:48 crc kubenswrapper[4719]: E1215 13:37:48.720945 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe\": container with ID starting with 16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe not found: ID does not exist" containerID="16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.720985 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe"} err="failed to get container status \"16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe\": rpc error: code = NotFound desc = could not find container \"16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe\": container with ID starting with 16315a77c484e8f2d321bf1fa496ff7eb070b97ffca8e648c43b4b67acdf50fe not found: ID does not exist" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.721011 4719 scope.go:117] "RemoveContainer" containerID="b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44" Dec 15 13:37:48 crc kubenswrapper[4719]: E1215 13:37:48.721239 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44\": container with ID starting with b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44 not found: ID does not exist" containerID="b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.721263 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44"} err="failed to get container status \"b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44\": rpc error: code = NotFound desc = could not find container \"b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44\": container with ID starting with b01cf0568e3a2203cd6c4969d9febda7aa2fbe0370eb7a5edb8a4eddd0cffe44 not found: ID does not exist" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.721280 4719 scope.go:117] "RemoveContainer" containerID="72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7" Dec 15 13:37:48 crc kubenswrapper[4719]: E1215 13:37:48.721487 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7\": container with ID starting with 72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7 not found: ID does not exist" containerID="72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7" Dec 15 13:37:48 crc kubenswrapper[4719]: I1215 13:37:48.721506 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7"} err="failed to get container status \"72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7\": rpc error: code = NotFound desc = could not find container \"72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7\": container with ID starting with 72ed792c392ca6dcf9cbb0c7c60d0e31c283e4d58895459c044987d346c2b5d7 not found: ID does not exist" Dec 15 13:37:49 crc kubenswrapper[4719]: I1215 13:37:49.485623 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e5797a5-9b7b-4613-8522-e223c3fe577a" path="/var/lib/kubelet/pods/3e5797a5-9b7b-4613-8522-e223c3fe577a/volumes" Dec 15 13:37:51 crc kubenswrapper[4719]: I1215 13:37:51.357441 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:37:51 crc kubenswrapper[4719]: I1215 13:37:51.357967 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:37:53 crc kubenswrapper[4719]: I1215 13:37:53.133482 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:53 crc kubenswrapper[4719]: I1215 13:37:53.159225 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:55 crc kubenswrapper[4719]: I1215 13:37:55.550575 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jqf7"] Dec 15 13:37:55 crc kubenswrapper[4719]: I1215 13:37:55.555806 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9jqf7" podUID="ddd690d7-223c-49c2-b743-79caec2762bb" containerName="registry-server" containerID="cri-o://d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425" gracePeriod=2 Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.129578 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.234257 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8tfl\" (UniqueName: \"kubernetes.io/projected/ddd690d7-223c-49c2-b743-79caec2762bb-kube-api-access-l8tfl\") pod \"ddd690d7-223c-49c2-b743-79caec2762bb\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.234394 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-catalog-content\") pod \"ddd690d7-223c-49c2-b743-79caec2762bb\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.234567 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-utilities\") pod \"ddd690d7-223c-49c2-b743-79caec2762bb\" (UID: \"ddd690d7-223c-49c2-b743-79caec2762bb\") " Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.235229 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-utilities" (OuterVolumeSpecName: "utilities") pod "ddd690d7-223c-49c2-b743-79caec2762bb" (UID: "ddd690d7-223c-49c2-b743-79caec2762bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.244077 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddd690d7-223c-49c2-b743-79caec2762bb-kube-api-access-l8tfl" (OuterVolumeSpecName: "kube-api-access-l8tfl") pod "ddd690d7-223c-49c2-b743-79caec2762bb" (UID: "ddd690d7-223c-49c2-b743-79caec2762bb"). InnerVolumeSpecName "kube-api-access-l8tfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.263208 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ddd690d7-223c-49c2-b743-79caec2762bb" (UID: "ddd690d7-223c-49c2-b743-79caec2762bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.337114 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8tfl\" (UniqueName: \"kubernetes.io/projected/ddd690d7-223c-49c2-b743-79caec2762bb-kube-api-access-l8tfl\") on node \"crc\" DevicePath \"\"" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.337459 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.337474 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddd690d7-223c-49c2-b743-79caec2762bb-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.534753 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t69mj"] Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.535181 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t69mj" podUID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerName="registry-server" containerID="cri-o://e2ab44da35b1f75d59fab94a314c1f17149fd244dd512d9e50f8ab7965121564" gracePeriod=2 Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.680381 4719 generic.go:334] "Generic (PLEG): container finished" podID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerID="e2ab44da35b1f75d59fab94a314c1f17149fd244dd512d9e50f8ab7965121564" exitCode=0 Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.680940 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t69mj" event={"ID":"01e39a53-c817-412f-bf0e-e5293c6b0a9f","Type":"ContainerDied","Data":"e2ab44da35b1f75d59fab94a314c1f17149fd244dd512d9e50f8ab7965121564"} Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.686821 4719 generic.go:334] "Generic (PLEG): container finished" podID="ddd690d7-223c-49c2-b743-79caec2762bb" containerID="d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425" exitCode=0 Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.686878 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jqf7" event={"ID":"ddd690d7-223c-49c2-b743-79caec2762bb","Type":"ContainerDied","Data":"d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425"} Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.686907 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jqf7" event={"ID":"ddd690d7-223c-49c2-b743-79caec2762bb","Type":"ContainerDied","Data":"556a14ab7eac595654ed7e301c4c6c5889c3a9629091e7b02568bb73a304f6f4"} Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.686924 4719 scope.go:117] "RemoveContainer" containerID="d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.687196 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jqf7" Dec 15 13:37:56 crc kubenswrapper[4719]: E1215 13:37:56.702874 4719 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01e39a53_c817_412f_bf0e_e5293c6b0a9f.slice/crio-e2ab44da35b1f75d59fab94a314c1f17149fd244dd512d9e50f8ab7965121564.scope\": RecentStats: unable to find data in memory cache]" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.716181 4719 scope.go:117] "RemoveContainer" containerID="5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.749099 4719 scope.go:117] "RemoveContainer" containerID="5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.756286 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jqf7"] Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.772090 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jqf7"] Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.835271 4719 scope.go:117] "RemoveContainer" containerID="d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425" Dec 15 13:37:56 crc kubenswrapper[4719]: E1215 13:37:56.839271 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425\": container with ID starting with d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425 not found: ID does not exist" containerID="d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.839312 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425"} err="failed to get container status \"d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425\": rpc error: code = NotFound desc = could not find container \"d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425\": container with ID starting with d2ef01f0e1853775cbdd5f34600e15d51b0337c69d615d62ff01df4ef5f1e425 not found: ID does not exist" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.839337 4719 scope.go:117] "RemoveContainer" containerID="5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b" Dec 15 13:37:56 crc kubenswrapper[4719]: E1215 13:37:56.839901 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b\": container with ID starting with 5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b not found: ID does not exist" containerID="5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.839943 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b"} err="failed to get container status \"5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b\": rpc error: code = NotFound desc = could not find container \"5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b\": container with ID starting with 5e98bb8f335f0d1232b060068dd975e8d1e36de0edbc3543a69e10c0c9f93d5b not found: ID does not exist" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.839972 4719 scope.go:117] "RemoveContainer" containerID="5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43" Dec 15 13:37:56 crc kubenswrapper[4719]: E1215 13:37:56.844508 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43\": container with ID starting with 5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43 not found: ID does not exist" containerID="5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43" Dec 15 13:37:56 crc kubenswrapper[4719]: I1215 13:37:56.844553 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43"} err="failed to get container status \"5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43\": rpc error: code = NotFound desc = could not find container \"5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43\": container with ID starting with 5922de7cde1d9bce610de2d8d87cdfd8d05fdc4e816efa1b86f277f9f3fb5a43 not found: ID does not exist" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.124847 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.258870 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-catalog-content\") pod \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.258936 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mkxv\" (UniqueName: \"kubernetes.io/projected/01e39a53-c817-412f-bf0e-e5293c6b0a9f-kube-api-access-8mkxv\") pod \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.259061 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-utilities\") pod \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\" (UID: \"01e39a53-c817-412f-bf0e-e5293c6b0a9f\") " Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.259888 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-utilities" (OuterVolumeSpecName: "utilities") pod "01e39a53-c817-412f-bf0e-e5293c6b0a9f" (UID: "01e39a53-c817-412f-bf0e-e5293c6b0a9f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.274074 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01e39a53-c817-412f-bf0e-e5293c6b0a9f-kube-api-access-8mkxv" (OuterVolumeSpecName: "kube-api-access-8mkxv") pod "01e39a53-c817-412f-bf0e-e5293c6b0a9f" (UID: "01e39a53-c817-412f-bf0e-e5293c6b0a9f"). InnerVolumeSpecName "kube-api-access-8mkxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.332443 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01e39a53-c817-412f-bf0e-e5293c6b0a9f" (UID: "01e39a53-c817-412f-bf0e-e5293c6b0a9f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.361127 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.361167 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mkxv\" (UniqueName: \"kubernetes.io/projected/01e39a53-c817-412f-bf0e-e5293c6b0a9f-kube-api-access-8mkxv\") on node \"crc\" DevicePath \"\"" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.361178 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01e39a53-c817-412f-bf0e-e5293c6b0a9f-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.486611 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddd690d7-223c-49c2-b743-79caec2762bb" path="/var/lib/kubelet/pods/ddd690d7-223c-49c2-b743-79caec2762bb/volumes" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.702157 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t69mj" event={"ID":"01e39a53-c817-412f-bf0e-e5293c6b0a9f","Type":"ContainerDied","Data":"0e1f21555e3dcedc6747bc040157826bf0527eec685636acd692ed8eec7e2835"} Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.702625 4719 scope.go:117] "RemoveContainer" containerID="e2ab44da35b1f75d59fab94a314c1f17149fd244dd512d9e50f8ab7965121564" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.702791 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t69mj" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.732929 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t69mj"] Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.738558 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t69mj"] Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.738704 4719 scope.go:117] "RemoveContainer" containerID="084f036166365287fa10de6ef747b24827f704b6f4cce3545f522183d5ba5c60" Dec 15 13:37:57 crc kubenswrapper[4719]: I1215 13:37:57.771564 4719 scope.go:117] "RemoveContainer" containerID="36ee25b082e1b00b8348038a9d74c4531f97e48b2a8e5bb01d4ddc1d985960c7" Dec 15 13:37:59 crc kubenswrapper[4719]: I1215 13:37:59.485590 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" path="/var/lib/kubelet/pods/01e39a53-c817-412f-bf0e-e5293c6b0a9f/volumes" Dec 15 13:38:06 crc kubenswrapper[4719]: I1215 13:38:06.528964 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-96g4h_49178e96-75a4-4b6d-8540-781182eb8123/kube-rbac-proxy/0.log" Dec 15 13:38:06 crc kubenswrapper[4719]: I1215 13:38:06.561541 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-96g4h_49178e96-75a4-4b6d-8540-781182eb8123/controller/0.log" Dec 15 13:38:06 crc kubenswrapper[4719]: I1215 13:38:06.680482 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-crztf_6bac73e3-46cd-4cca-bb26-1ae16f1c804e/frr-k8s-webhook-server/0.log" Dec 15 13:38:06 crc kubenswrapper[4719]: I1215 13:38:06.817007 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-frr-files/0.log" Dec 15 13:38:07 crc kubenswrapper[4719]: I1215 13:38:07.253802 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-reloader/0.log" Dec 15 13:38:07 crc kubenswrapper[4719]: I1215 13:38:07.301337 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-frr-files/0.log" Dec 15 13:38:07 crc kubenswrapper[4719]: I1215 13:38:07.378117 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-metrics/0.log" Dec 15 13:38:07 crc kubenswrapper[4719]: I1215 13:38:07.398319 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-reloader/0.log" Dec 15 13:38:07 crc kubenswrapper[4719]: I1215 13:38:07.626837 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-frr-files/0.log" Dec 15 13:38:07 crc kubenswrapper[4719]: I1215 13:38:07.669840 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-metrics/0.log" Dec 15 13:38:07 crc kubenswrapper[4719]: I1215 13:38:07.718039 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-reloader/0.log" Dec 15 13:38:07 crc kubenswrapper[4719]: I1215 13:38:07.740214 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-metrics/0.log" Dec 15 13:38:07 crc kubenswrapper[4719]: I1215 13:38:07.958091 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-reloader/0.log" Dec 15 13:38:08 crc kubenswrapper[4719]: I1215 13:38:08.003263 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-metrics/0.log" Dec 15 13:38:08 crc kubenswrapper[4719]: I1215 13:38:08.032543 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-frr-files/0.log" Dec 15 13:38:08 crc kubenswrapper[4719]: I1215 13:38:08.037537 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/controller/0.log" Dec 15 13:38:08 crc kubenswrapper[4719]: I1215 13:38:08.198062 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/frr-metrics/0.log" Dec 15 13:38:08 crc kubenswrapper[4719]: I1215 13:38:08.233569 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/kube-rbac-proxy-frr/0.log" Dec 15 13:38:08 crc kubenswrapper[4719]: I1215 13:38:08.322121 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/kube-rbac-proxy/0.log" Dec 15 13:38:08 crc kubenswrapper[4719]: I1215 13:38:08.507531 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/reloader/0.log" Dec 15 13:38:08 crc kubenswrapper[4719]: I1215 13:38:08.638794 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6cf5cc9b68-dpxck_5009ea4d-5e13-4ac3-ad25-8c0976b0aafb/manager/0.log" Dec 15 13:38:09 crc kubenswrapper[4719]: I1215 13:38:09.285155 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-76fbf99b47-d626r_15904dfe-b1ca-4922-95cf-6f3724fae181/webhook-server/0.log" Dec 15 13:38:09 crc kubenswrapper[4719]: I1215 13:38:09.444320 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tdxf5_72e46a3d-9b79-404d-80ba-68a363c23b2b/kube-rbac-proxy/0.log" Dec 15 13:38:09 crc kubenswrapper[4719]: I1215 13:38:09.712475 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/frr/0.log" Dec 15 13:38:10 crc kubenswrapper[4719]: I1215 13:38:10.076948 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tdxf5_72e46a3d-9b79-404d-80ba-68a363c23b2b/speaker/0.log" Dec 15 13:38:21 crc kubenswrapper[4719]: I1215 13:38:21.356679 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:38:21 crc kubenswrapper[4719]: I1215 13:38:21.357200 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:38:24 crc kubenswrapper[4719]: I1215 13:38:24.260973 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/util/0.log" Dec 15 13:38:24 crc kubenswrapper[4719]: I1215 13:38:24.469454 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/pull/0.log" Dec 15 13:38:24 crc kubenswrapper[4719]: I1215 13:38:24.473549 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/pull/0.log" Dec 15 13:38:24 crc kubenswrapper[4719]: I1215 13:38:24.508564 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/util/0.log" Dec 15 13:38:24 crc kubenswrapper[4719]: I1215 13:38:24.790232 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/util/0.log" Dec 15 13:38:24 crc kubenswrapper[4719]: I1215 13:38:24.791231 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/extract/0.log" Dec 15 13:38:24 crc kubenswrapper[4719]: I1215 13:38:24.814255 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/pull/0.log" Dec 15 13:38:25 crc kubenswrapper[4719]: I1215 13:38:25.008751 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/util/0.log" Dec 15 13:38:25 crc kubenswrapper[4719]: I1215 13:38:25.205371 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/util/0.log" Dec 15 13:38:25 crc kubenswrapper[4719]: I1215 13:38:25.231385 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/pull/0.log" Dec 15 13:38:25 crc kubenswrapper[4719]: I1215 13:38:25.235044 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/pull/0.log" Dec 15 13:38:25 crc kubenswrapper[4719]: I1215 13:38:25.481068 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/pull/0.log" Dec 15 13:38:25 crc kubenswrapper[4719]: I1215 13:38:25.533022 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/util/0.log" Dec 15 13:38:25 crc kubenswrapper[4719]: I1215 13:38:25.561262 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/extract/0.log" Dec 15 13:38:25 crc kubenswrapper[4719]: I1215 13:38:25.772451 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-utilities/0.log" Dec 15 13:38:25 crc kubenswrapper[4719]: I1215 13:38:25.987868 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-content/0.log" Dec 15 13:38:26 crc kubenswrapper[4719]: I1215 13:38:26.010223 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-utilities/0.log" Dec 15 13:38:26 crc kubenswrapper[4719]: I1215 13:38:26.041299 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-content/0.log" Dec 15 13:38:26 crc kubenswrapper[4719]: I1215 13:38:26.256531 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-utilities/0.log" Dec 15 13:38:26 crc kubenswrapper[4719]: I1215 13:38:26.284359 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-content/0.log" Dec 15 13:38:26 crc kubenswrapper[4719]: I1215 13:38:26.564109 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-utilities/0.log" Dec 15 13:38:26 crc kubenswrapper[4719]: I1215 13:38:26.946655 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-utilities/0.log" Dec 15 13:38:27 crc kubenswrapper[4719]: I1215 13:38:27.024322 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-content/0.log" Dec 15 13:38:27 crc kubenswrapper[4719]: I1215 13:38:27.043156 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/registry-server/0.log" Dec 15 13:38:27 crc kubenswrapper[4719]: I1215 13:38:27.074435 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-content/0.log" Dec 15 13:38:27 crc kubenswrapper[4719]: I1215 13:38:27.995689 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-utilities/0.log" Dec 15 13:38:28 crc kubenswrapper[4719]: I1215 13:38:28.014035 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-content/0.log" Dec 15 13:38:28 crc kubenswrapper[4719]: I1215 13:38:28.390792 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kxrn8_e804e02c-d21b-4881-9bc8-2de04a0928f9/marketplace-operator/0.log" Dec 15 13:38:28 crc kubenswrapper[4719]: I1215 13:38:28.516044 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-utilities/0.log" Dec 15 13:38:28 crc kubenswrapper[4719]: I1215 13:38:28.876888 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-content/0.log" Dec 15 13:38:28 crc kubenswrapper[4719]: I1215 13:38:28.930718 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/registry-server/0.log" Dec 15 13:38:28 crc kubenswrapper[4719]: I1215 13:38:28.937781 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-utilities/0.log" Dec 15 13:38:28 crc kubenswrapper[4719]: I1215 13:38:28.967915 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-content/0.log" Dec 15 13:38:29 crc kubenswrapper[4719]: I1215 13:38:29.602522 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-content/0.log" Dec 15 13:38:29 crc kubenswrapper[4719]: I1215 13:38:29.641996 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-utilities/0.log" Dec 15 13:38:29 crc kubenswrapper[4719]: I1215 13:38:29.649457 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-utilities/0.log" Dec 15 13:38:29 crc kubenswrapper[4719]: I1215 13:38:29.695556 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/registry-server/0.log" Dec 15 13:38:29 crc kubenswrapper[4719]: I1215 13:38:29.874568 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-content/0.log" Dec 15 13:38:29 crc kubenswrapper[4719]: I1215 13:38:29.898133 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-content/0.log" Dec 15 13:38:29 crc kubenswrapper[4719]: I1215 13:38:29.917123 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-utilities/0.log" Dec 15 13:38:30 crc kubenswrapper[4719]: I1215 13:38:30.194756 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-utilities/0.log" Dec 15 13:38:30 crc kubenswrapper[4719]: I1215 13:38:30.213592 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-content/0.log" Dec 15 13:38:30 crc kubenswrapper[4719]: I1215 13:38:30.497863 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/registry-server/0.log" Dec 15 13:38:51 crc kubenswrapper[4719]: I1215 13:38:51.356903 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:38:51 crc kubenswrapper[4719]: I1215 13:38:51.357340 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:38:51 crc kubenswrapper[4719]: I1215 13:38:51.357385 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 13:38:51 crc kubenswrapper[4719]: I1215 13:38:51.358086 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e5f68797b715b1014a038536c351a5cc8510f4641ca0aafa31da234a3cd21cac"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 13:38:51 crc kubenswrapper[4719]: I1215 13:38:51.358135 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://e5f68797b715b1014a038536c351a5cc8510f4641ca0aafa31da234a3cd21cac" gracePeriod=600 Dec 15 13:38:52 crc kubenswrapper[4719]: I1215 13:38:52.141597 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="e5f68797b715b1014a038536c351a5cc8510f4641ca0aafa31da234a3cd21cac" exitCode=0 Dec 15 13:38:52 crc kubenswrapper[4719]: I1215 13:38:52.141654 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"e5f68797b715b1014a038536c351a5cc8510f4641ca0aafa31da234a3cd21cac"} Dec 15 13:38:52 crc kubenswrapper[4719]: I1215 13:38:52.141914 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a"} Dec 15 13:38:52 crc kubenswrapper[4719]: I1215 13:38:52.141937 4719 scope.go:117] "RemoveContainer" containerID="310746bab9798f8409c40382afe9cd0a15be4a51a72f23b7fe6764af7f4c94b7" Dec 15 13:40:51 crc kubenswrapper[4719]: I1215 13:40:51.356581 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:40:51 crc kubenswrapper[4719]: I1215 13:40:51.357151 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:40:57 crc kubenswrapper[4719]: I1215 13:40:57.224870 4719 generic.go:334] "Generic (PLEG): container finished" podID="2397d720-99e1-4da0-93ff-aac39122fadf" containerID="2a69ac7487a338ebca56a28f9b35c53d4a63471163932223e8deebab40507a08" exitCode=0 Dec 15 13:40:57 crc kubenswrapper[4719]: I1215 13:40:57.224891 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8s97/must-gather-88wmh" event={"ID":"2397d720-99e1-4da0-93ff-aac39122fadf","Type":"ContainerDied","Data":"2a69ac7487a338ebca56a28f9b35c53d4a63471163932223e8deebab40507a08"} Dec 15 13:40:57 crc kubenswrapper[4719]: I1215 13:40:57.226268 4719 scope.go:117] "RemoveContainer" containerID="2a69ac7487a338ebca56a28f9b35c53d4a63471163932223e8deebab40507a08" Dec 15 13:40:57 crc kubenswrapper[4719]: I1215 13:40:57.338898 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c8s97_must-gather-88wmh_2397d720-99e1-4da0-93ff-aac39122fadf/gather/0.log" Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.100217 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c8s97/must-gather-88wmh"] Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.101134 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-c8s97/must-gather-88wmh" podUID="2397d720-99e1-4da0-93ff-aac39122fadf" containerName="copy" containerID="cri-o://a7c34f9fe63109ec0bdf8bf2b6af765b24d417a1f9b30bbf9d89b9249f3103d7" gracePeriod=2 Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.109130 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c8s97/must-gather-88wmh"] Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.320708 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c8s97_must-gather-88wmh_2397d720-99e1-4da0-93ff-aac39122fadf/copy/0.log" Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.321520 4719 generic.go:334] "Generic (PLEG): container finished" podID="2397d720-99e1-4da0-93ff-aac39122fadf" containerID="a7c34f9fe63109ec0bdf8bf2b6af765b24d417a1f9b30bbf9d89b9249f3103d7" exitCode=143 Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.630131 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c8s97_must-gather-88wmh_2397d720-99e1-4da0-93ff-aac39122fadf/copy/0.log" Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.630919 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/must-gather-88wmh" Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.782834 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2397d720-99e1-4da0-93ff-aac39122fadf-must-gather-output\") pod \"2397d720-99e1-4da0-93ff-aac39122fadf\" (UID: \"2397d720-99e1-4da0-93ff-aac39122fadf\") " Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.783054 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcj72\" (UniqueName: \"kubernetes.io/projected/2397d720-99e1-4da0-93ff-aac39122fadf-kube-api-access-dcj72\") pod \"2397d720-99e1-4da0-93ff-aac39122fadf\" (UID: \"2397d720-99e1-4da0-93ff-aac39122fadf\") " Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.965795 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2397d720-99e1-4da0-93ff-aac39122fadf-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2397d720-99e1-4da0-93ff-aac39122fadf" (UID: "2397d720-99e1-4da0-93ff-aac39122fadf"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:41:06 crc kubenswrapper[4719]: I1215 13:41:06.987092 4719 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2397d720-99e1-4da0-93ff-aac39122fadf-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 15 13:41:07 crc kubenswrapper[4719]: I1215 13:41:07.292007 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2397d720-99e1-4da0-93ff-aac39122fadf-kube-api-access-dcj72" (OuterVolumeSpecName: "kube-api-access-dcj72") pod "2397d720-99e1-4da0-93ff-aac39122fadf" (UID: "2397d720-99e1-4da0-93ff-aac39122fadf"). InnerVolumeSpecName "kube-api-access-dcj72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:41:07 crc kubenswrapper[4719]: I1215 13:41:07.293682 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcj72\" (UniqueName: \"kubernetes.io/projected/2397d720-99e1-4da0-93ff-aac39122fadf-kube-api-access-dcj72\") on node \"crc\" DevicePath \"\"" Dec 15 13:41:07 crc kubenswrapper[4719]: I1215 13:41:07.330969 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c8s97_must-gather-88wmh_2397d720-99e1-4da0-93ff-aac39122fadf/copy/0.log" Dec 15 13:41:07 crc kubenswrapper[4719]: I1215 13:41:07.332249 4719 scope.go:117] "RemoveContainer" containerID="a7c34f9fe63109ec0bdf8bf2b6af765b24d417a1f9b30bbf9d89b9249f3103d7" Dec 15 13:41:07 crc kubenswrapper[4719]: I1215 13:41:07.332503 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8s97/must-gather-88wmh" Dec 15 13:41:07 crc kubenswrapper[4719]: I1215 13:41:07.353144 4719 scope.go:117] "RemoveContainer" containerID="2a69ac7487a338ebca56a28f9b35c53d4a63471163932223e8deebab40507a08" Dec 15 13:41:07 crc kubenswrapper[4719]: I1215 13:41:07.486007 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2397d720-99e1-4da0-93ff-aac39122fadf" path="/var/lib/kubelet/pods/2397d720-99e1-4da0-93ff-aac39122fadf/volumes" Dec 15 13:41:21 crc kubenswrapper[4719]: I1215 13:41:21.356791 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:41:21 crc kubenswrapper[4719]: I1215 13:41:21.357487 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:41:36 crc kubenswrapper[4719]: I1215 13:41:36.087184 4719 scope.go:117] "RemoveContainer" containerID="2001c130490d9811c3fec4ce675a4bdf200ee83ddc695fb8475c936662d5d1bb" Dec 15 13:41:51 crc kubenswrapper[4719]: I1215 13:41:51.356636 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:41:51 crc kubenswrapper[4719]: I1215 13:41:51.357127 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:41:51 crc kubenswrapper[4719]: I1215 13:41:51.357176 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 13:41:51 crc kubenswrapper[4719]: I1215 13:41:51.357816 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 13:41:51 crc kubenswrapper[4719]: I1215 13:41:51.357886 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" gracePeriod=600 Dec 15 13:41:51 crc kubenswrapper[4719]: I1215 13:41:51.719344 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" exitCode=0 Dec 15 13:41:51 crc kubenswrapper[4719]: I1215 13:41:51.719427 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a"} Dec 15 13:41:51 crc kubenswrapper[4719]: I1215 13:41:51.719683 4719 scope.go:117] "RemoveContainer" containerID="e5f68797b715b1014a038536c351a5cc8510f4641ca0aafa31da234a3cd21cac" Dec 15 13:41:52 crc kubenswrapper[4719]: E1215 13:41:52.125705 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:41:52 crc kubenswrapper[4719]: I1215 13:41:52.730063 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:41:52 crc kubenswrapper[4719]: E1215 13:41:52.730631 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:42:06 crc kubenswrapper[4719]: I1215 13:42:06.475896 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:42:06 crc kubenswrapper[4719]: E1215 13:42:06.476640 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:42:17 crc kubenswrapper[4719]: I1215 13:42:17.476187 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:42:17 crc kubenswrapper[4719]: E1215 13:42:17.476990 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:42:30 crc kubenswrapper[4719]: I1215 13:42:30.477003 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:42:30 crc kubenswrapper[4719]: E1215 13:42:30.477745 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:42:36 crc kubenswrapper[4719]: I1215 13:42:36.252529 4719 scope.go:117] "RemoveContainer" containerID="92b61a7e2674c517207964a2c542c7e81cd9a155997937f0d41cd8c5c4ae5e81" Dec 15 13:42:44 crc kubenswrapper[4719]: I1215 13:42:44.476543 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:42:44 crc kubenswrapper[4719]: E1215 13:42:44.477577 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:42:57 crc kubenswrapper[4719]: I1215 13:42:57.475746 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:42:57 crc kubenswrapper[4719]: E1215 13:42:57.476554 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:43:11 crc kubenswrapper[4719]: I1215 13:43:11.475988 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:43:11 crc kubenswrapper[4719]: E1215 13:43:11.476757 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:43:22 crc kubenswrapper[4719]: I1215 13:43:22.476295 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:43:22 crc kubenswrapper[4719]: E1215 13:43:22.477277 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:43:34 crc kubenswrapper[4719]: I1215 13:43:34.476148 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:43:34 crc kubenswrapper[4719]: E1215 13:43:34.477059 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:43:48 crc kubenswrapper[4719]: I1215 13:43:48.476407 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:43:48 crc kubenswrapper[4719]: E1215 13:43:48.478005 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.973458 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lfsjs"] Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974304 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd690d7-223c-49c2-b743-79caec2762bb" containerName="registry-server" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974319 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd690d7-223c-49c2-b743-79caec2762bb" containerName="registry-server" Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974331 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerName="registry-server" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974339 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerName="registry-server" Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974354 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerName="extract-utilities" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974361 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerName="extract-utilities" Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974375 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd690d7-223c-49c2-b743-79caec2762bb" containerName="extract-utilities" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974381 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd690d7-223c-49c2-b743-79caec2762bb" containerName="extract-utilities" Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974395 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerName="extract-content" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974402 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerName="extract-content" Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974421 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2397d720-99e1-4da0-93ff-aac39122fadf" containerName="gather" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974427 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2397d720-99e1-4da0-93ff-aac39122fadf" containerName="gather" Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974438 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerName="registry-server" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974444 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerName="registry-server" Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974456 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2397d720-99e1-4da0-93ff-aac39122fadf" containerName="copy" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974462 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="2397d720-99e1-4da0-93ff-aac39122fadf" containerName="copy" Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974478 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerName="extract-utilities" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974484 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerName="extract-utilities" Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974503 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd690d7-223c-49c2-b743-79caec2762bb" containerName="extract-content" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974509 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd690d7-223c-49c2-b743-79caec2762bb" containerName="extract-content" Dec 15 13:43:52 crc kubenswrapper[4719]: E1215 13:43:52.974523 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerName="extract-content" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974528 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerName="extract-content" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974687 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e39a53-c817-412f-bf0e-e5293c6b0a9f" containerName="registry-server" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974698 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="2397d720-99e1-4da0-93ff-aac39122fadf" containerName="gather" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974711 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="2397d720-99e1-4da0-93ff-aac39122fadf" containerName="copy" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974719 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e5797a5-9b7b-4613-8522-e223c3fe577a" containerName="registry-server" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.974736 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddd690d7-223c-49c2-b743-79caec2762bb" containerName="registry-server" Dec 15 13:43:52 crc kubenswrapper[4719]: I1215 13:43:52.976793 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.032303 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lfsjs"] Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.125653 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-utilities\") pod \"redhat-operators-lfsjs\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.125749 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcwp5\" (UniqueName: \"kubernetes.io/projected/6a8b0687-7b15-4702-8ed7-6a52c51679d9-kube-api-access-lcwp5\") pod \"redhat-operators-lfsjs\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.125823 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-catalog-content\") pod \"redhat-operators-lfsjs\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.228144 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-utilities\") pod \"redhat-operators-lfsjs\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.228450 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcwp5\" (UniqueName: \"kubernetes.io/projected/6a8b0687-7b15-4702-8ed7-6a52c51679d9-kube-api-access-lcwp5\") pod \"redhat-operators-lfsjs\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.228583 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-catalog-content\") pod \"redhat-operators-lfsjs\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.228634 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-utilities\") pod \"redhat-operators-lfsjs\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.228873 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-catalog-content\") pod \"redhat-operators-lfsjs\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.263052 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcwp5\" (UniqueName: \"kubernetes.io/projected/6a8b0687-7b15-4702-8ed7-6a52c51679d9-kube-api-access-lcwp5\") pod \"redhat-operators-lfsjs\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.294118 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:43:53 crc kubenswrapper[4719]: I1215 13:43:53.785785 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lfsjs"] Dec 15 13:43:54 crc kubenswrapper[4719]: I1215 13:43:54.753249 4719 generic.go:334] "Generic (PLEG): container finished" podID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerID="5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da" exitCode=0 Dec 15 13:43:54 crc kubenswrapper[4719]: I1215 13:43:54.753448 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfsjs" event={"ID":"6a8b0687-7b15-4702-8ed7-6a52c51679d9","Type":"ContainerDied","Data":"5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da"} Dec 15 13:43:54 crc kubenswrapper[4719]: I1215 13:43:54.753724 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfsjs" event={"ID":"6a8b0687-7b15-4702-8ed7-6a52c51679d9","Type":"ContainerStarted","Data":"0bc514f1afab207dc107c3bed000073b019a6ec192ca566b7d16a6d8193dc502"} Dec 15 13:43:54 crc kubenswrapper[4719]: I1215 13:43:54.761742 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 13:43:55 crc kubenswrapper[4719]: I1215 13:43:55.766119 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfsjs" event={"ID":"6a8b0687-7b15-4702-8ed7-6a52c51679d9","Type":"ContainerStarted","Data":"e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f"} Dec 15 13:43:59 crc kubenswrapper[4719]: I1215 13:43:59.494112 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:43:59 crc kubenswrapper[4719]: E1215 13:43:59.494924 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:43:59 crc kubenswrapper[4719]: I1215 13:43:59.798914 4719 generic.go:334] "Generic (PLEG): container finished" podID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerID="e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f" exitCode=0 Dec 15 13:43:59 crc kubenswrapper[4719]: I1215 13:43:59.798954 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfsjs" event={"ID":"6a8b0687-7b15-4702-8ed7-6a52c51679d9","Type":"ContainerDied","Data":"e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f"} Dec 15 13:44:01 crc kubenswrapper[4719]: I1215 13:44:01.820815 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfsjs" event={"ID":"6a8b0687-7b15-4702-8ed7-6a52c51679d9","Type":"ContainerStarted","Data":"d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3"} Dec 15 13:44:01 crc kubenswrapper[4719]: I1215 13:44:01.849174 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lfsjs" podStartSLOduration=3.753514818 podStartE2EDuration="9.84915574s" podCreationTimestamp="2025-12-15 13:43:52 +0000 UTC" firstStartedPulling="2025-12-15 13:43:54.76133148 +0000 UTC m=+5195.703624520" lastFinishedPulling="2025-12-15 13:44:00.856972402 +0000 UTC m=+5201.799265442" observedRunningTime="2025-12-15 13:44:01.844303977 +0000 UTC m=+5202.786597027" watchObservedRunningTime="2025-12-15 13:44:01.84915574 +0000 UTC m=+5202.791448770" Dec 15 13:44:03 crc kubenswrapper[4719]: I1215 13:44:03.294887 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:44:03 crc kubenswrapper[4719]: I1215 13:44:03.295276 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:44:04 crc kubenswrapper[4719]: I1215 13:44:04.339034 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lfsjs" podUID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerName="registry-server" probeResult="failure" output=< Dec 15 13:44:04 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 13:44:04 crc kubenswrapper[4719]: > Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.714576 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bf8hx/must-gather-xzsl8"] Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.722834 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/must-gather-xzsl8" Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.731411 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bf8hx"/"openshift-service-ca.crt" Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.731592 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bf8hx"/"kube-root-ca.crt" Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.734906 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-bf8hx"/"default-dockercfg-w9kd8" Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.748967 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bf8hx/must-gather-xzsl8"] Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.835612 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f3679f93-2fdb-4155-be86-eac3670559ea-must-gather-output\") pod \"must-gather-xzsl8\" (UID: \"f3679f93-2fdb-4155-be86-eac3670559ea\") " pod="openshift-must-gather-bf8hx/must-gather-xzsl8" Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.835669 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6mp8\" (UniqueName: \"kubernetes.io/projected/f3679f93-2fdb-4155-be86-eac3670559ea-kube-api-access-f6mp8\") pod \"must-gather-xzsl8\" (UID: \"f3679f93-2fdb-4155-be86-eac3670559ea\") " pod="openshift-must-gather-bf8hx/must-gather-xzsl8" Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.938436 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f3679f93-2fdb-4155-be86-eac3670559ea-must-gather-output\") pod \"must-gather-xzsl8\" (UID: \"f3679f93-2fdb-4155-be86-eac3670559ea\") " pod="openshift-must-gather-bf8hx/must-gather-xzsl8" Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.938495 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6mp8\" (UniqueName: \"kubernetes.io/projected/f3679f93-2fdb-4155-be86-eac3670559ea-kube-api-access-f6mp8\") pod \"must-gather-xzsl8\" (UID: \"f3679f93-2fdb-4155-be86-eac3670559ea\") " pod="openshift-must-gather-bf8hx/must-gather-xzsl8" Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.938945 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f3679f93-2fdb-4155-be86-eac3670559ea-must-gather-output\") pod \"must-gather-xzsl8\" (UID: \"f3679f93-2fdb-4155-be86-eac3670559ea\") " pod="openshift-must-gather-bf8hx/must-gather-xzsl8" Dec 15 13:44:08 crc kubenswrapper[4719]: I1215 13:44:08.962744 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6mp8\" (UniqueName: \"kubernetes.io/projected/f3679f93-2fdb-4155-be86-eac3670559ea-kube-api-access-f6mp8\") pod \"must-gather-xzsl8\" (UID: \"f3679f93-2fdb-4155-be86-eac3670559ea\") " pod="openshift-must-gather-bf8hx/must-gather-xzsl8" Dec 15 13:44:09 crc kubenswrapper[4719]: I1215 13:44:09.051350 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/must-gather-xzsl8" Dec 15 13:44:09 crc kubenswrapper[4719]: I1215 13:44:09.517609 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bf8hx/must-gather-xzsl8"] Dec 15 13:44:09 crc kubenswrapper[4719]: I1215 13:44:09.894388 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/must-gather-xzsl8" event={"ID":"f3679f93-2fdb-4155-be86-eac3670559ea","Type":"ContainerStarted","Data":"f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648"} Dec 15 13:44:09 crc kubenswrapper[4719]: I1215 13:44:09.894946 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/must-gather-xzsl8" event={"ID":"f3679f93-2fdb-4155-be86-eac3670559ea","Type":"ContainerStarted","Data":"1face8c13ab24b2902518cce1d6c78184d79f75e07a01a26c3140620ccfbd857"} Dec 15 13:44:10 crc kubenswrapper[4719]: I1215 13:44:10.910339 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/must-gather-xzsl8" event={"ID":"f3679f93-2fdb-4155-be86-eac3670559ea","Type":"ContainerStarted","Data":"df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c"} Dec 15 13:44:10 crc kubenswrapper[4719]: I1215 13:44:10.936296 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bf8hx/must-gather-xzsl8" podStartSLOduration=2.936276616 podStartE2EDuration="2.936276616s" podCreationTimestamp="2025-12-15 13:44:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 13:44:10.927690475 +0000 UTC m=+5211.869983515" watchObservedRunningTime="2025-12-15 13:44:10.936276616 +0000 UTC m=+5211.878569646" Dec 15 13:44:13 crc kubenswrapper[4719]: I1215 13:44:13.355923 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:44:13 crc kubenswrapper[4719]: I1215 13:44:13.413922 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:44:13 crc kubenswrapper[4719]: I1215 13:44:13.605348 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lfsjs"] Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.188231 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bf8hx/crc-debug-xjw9p"] Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.189532 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.249389 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hxv5\" (UniqueName: \"kubernetes.io/projected/932e9f4f-4b8c-4403-aa58-feb7af89e510-kube-api-access-4hxv5\") pod \"crc-debug-xjw9p\" (UID: \"932e9f4f-4b8c-4403-aa58-feb7af89e510\") " pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.249660 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/932e9f4f-4b8c-4403-aa58-feb7af89e510-host\") pod \"crc-debug-xjw9p\" (UID: \"932e9f4f-4b8c-4403-aa58-feb7af89e510\") " pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.351422 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hxv5\" (UniqueName: \"kubernetes.io/projected/932e9f4f-4b8c-4403-aa58-feb7af89e510-kube-api-access-4hxv5\") pod \"crc-debug-xjw9p\" (UID: \"932e9f4f-4b8c-4403-aa58-feb7af89e510\") " pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.351472 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/932e9f4f-4b8c-4403-aa58-feb7af89e510-host\") pod \"crc-debug-xjw9p\" (UID: \"932e9f4f-4b8c-4403-aa58-feb7af89e510\") " pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.351636 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/932e9f4f-4b8c-4403-aa58-feb7af89e510-host\") pod \"crc-debug-xjw9p\" (UID: \"932e9f4f-4b8c-4403-aa58-feb7af89e510\") " pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.382590 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hxv5\" (UniqueName: \"kubernetes.io/projected/932e9f4f-4b8c-4403-aa58-feb7af89e510-kube-api-access-4hxv5\") pod \"crc-debug-xjw9p\" (UID: \"932e9f4f-4b8c-4403-aa58-feb7af89e510\") " pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.476302 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:44:14 crc kubenswrapper[4719]: E1215 13:44:14.476571 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.506410 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" Dec 15 13:44:14 crc kubenswrapper[4719]: W1215 13:44:14.536458 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod932e9f4f_4b8c_4403_aa58_feb7af89e510.slice/crio-b5b9c2169f19cd989d5f3b33736da0c2de0da3479fbde885404dec2b2da68bde WatchSource:0}: Error finding container b5b9c2169f19cd989d5f3b33736da0c2de0da3479fbde885404dec2b2da68bde: Status 404 returned error can't find the container with id b5b9c2169f19cd989d5f3b33736da0c2de0da3479fbde885404dec2b2da68bde Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.945350 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" event={"ID":"932e9f4f-4b8c-4403-aa58-feb7af89e510","Type":"ContainerStarted","Data":"e6f4b8fe43b6aa2fe7598b8bfe4e3bd03ce42ecefeca721c81a3debc70eb43fb"} Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.945928 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" event={"ID":"932e9f4f-4b8c-4403-aa58-feb7af89e510","Type":"ContainerStarted","Data":"b5b9c2169f19cd989d5f3b33736da0c2de0da3479fbde885404dec2b2da68bde"} Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.945525 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lfsjs" podUID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerName="registry-server" containerID="cri-o://d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3" gracePeriod=2 Dec 15 13:44:14 crc kubenswrapper[4719]: I1215 13:44:14.971366 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" podStartSLOduration=0.971346 podStartE2EDuration="971.346ms" podCreationTimestamp="2025-12-15 13:44:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-15 13:44:14.963703699 +0000 UTC m=+5215.905996729" watchObservedRunningTime="2025-12-15 13:44:14.971346 +0000 UTC m=+5215.913639030" Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.626684 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.675391 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcwp5\" (UniqueName: \"kubernetes.io/projected/6a8b0687-7b15-4702-8ed7-6a52c51679d9-kube-api-access-lcwp5\") pod \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.675515 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-utilities\") pod \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.675627 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-catalog-content\") pod \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\" (UID: \"6a8b0687-7b15-4702-8ed7-6a52c51679d9\") " Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.678556 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-utilities" (OuterVolumeSpecName: "utilities") pod "6a8b0687-7b15-4702-8ed7-6a52c51679d9" (UID: "6a8b0687-7b15-4702-8ed7-6a52c51679d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.684023 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a8b0687-7b15-4702-8ed7-6a52c51679d9-kube-api-access-lcwp5" (OuterVolumeSpecName: "kube-api-access-lcwp5") pod "6a8b0687-7b15-4702-8ed7-6a52c51679d9" (UID: "6a8b0687-7b15-4702-8ed7-6a52c51679d9"). InnerVolumeSpecName "kube-api-access-lcwp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.777541 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcwp5\" (UniqueName: \"kubernetes.io/projected/6a8b0687-7b15-4702-8ed7-6a52c51679d9-kube-api-access-lcwp5\") on node \"crc\" DevicePath \"\"" Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.777744 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.872425 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a8b0687-7b15-4702-8ed7-6a52c51679d9" (UID: "6a8b0687-7b15-4702-8ed7-6a52c51679d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.883991 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8b0687-7b15-4702-8ed7-6a52c51679d9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.961710 4719 generic.go:334] "Generic (PLEG): container finished" podID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerID="d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3" exitCode=0 Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.961759 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lfsjs" Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.961781 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfsjs" event={"ID":"6a8b0687-7b15-4702-8ed7-6a52c51679d9","Type":"ContainerDied","Data":"d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3"} Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.962209 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfsjs" event={"ID":"6a8b0687-7b15-4702-8ed7-6a52c51679d9","Type":"ContainerDied","Data":"0bc514f1afab207dc107c3bed000073b019a6ec192ca566b7d16a6d8193dc502"} Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.962294 4719 scope.go:117] "RemoveContainer" containerID="d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3" Dec 15 13:44:15 crc kubenswrapper[4719]: I1215 13:44:15.988929 4719 scope.go:117] "RemoveContainer" containerID="e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f" Dec 15 13:44:16 crc kubenswrapper[4719]: I1215 13:44:16.017131 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lfsjs"] Dec 15 13:44:16 crc kubenswrapper[4719]: I1215 13:44:16.028523 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lfsjs"] Dec 15 13:44:16 crc kubenswrapper[4719]: I1215 13:44:16.043598 4719 scope.go:117] "RemoveContainer" containerID="5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da" Dec 15 13:44:16 crc kubenswrapper[4719]: I1215 13:44:16.094487 4719 scope.go:117] "RemoveContainer" containerID="d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3" Dec 15 13:44:16 crc kubenswrapper[4719]: E1215 13:44:16.095956 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3\": container with ID starting with d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3 not found: ID does not exist" containerID="d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3" Dec 15 13:44:16 crc kubenswrapper[4719]: I1215 13:44:16.096010 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3"} err="failed to get container status \"d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3\": rpc error: code = NotFound desc = could not find container \"d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3\": container with ID starting with d0e9038c8dc162b9038c1df4284eb01d935789365ec1f0582e9c90bbaa89fde3 not found: ID does not exist" Dec 15 13:44:16 crc kubenswrapper[4719]: I1215 13:44:16.096038 4719 scope.go:117] "RemoveContainer" containerID="e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f" Dec 15 13:44:16 crc kubenswrapper[4719]: E1215 13:44:16.097062 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f\": container with ID starting with e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f not found: ID does not exist" containerID="e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f" Dec 15 13:44:16 crc kubenswrapper[4719]: I1215 13:44:16.097154 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f"} err="failed to get container status \"e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f\": rpc error: code = NotFound desc = could not find container \"e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f\": container with ID starting with e1ee513af637c925074fe6b1b4164dbbfe7d9b4989cf9557354db85469740a7f not found: ID does not exist" Dec 15 13:44:16 crc kubenswrapper[4719]: I1215 13:44:16.097168 4719 scope.go:117] "RemoveContainer" containerID="5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da" Dec 15 13:44:16 crc kubenswrapper[4719]: E1215 13:44:16.097492 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da\": container with ID starting with 5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da not found: ID does not exist" containerID="5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da" Dec 15 13:44:16 crc kubenswrapper[4719]: I1215 13:44:16.097511 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da"} err="failed to get container status \"5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da\": rpc error: code = NotFound desc = could not find container \"5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da\": container with ID starting with 5e8b94ad0f147a5d4cedf18853e0db8c558f6b0ce234b047713b21bf4acb54da not found: ID does not exist" Dec 15 13:44:17 crc kubenswrapper[4719]: I1215 13:44:17.486252 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" path="/var/lib/kubelet/pods/6a8b0687-7b15-4702-8ed7-6a52c51679d9/volumes" Dec 15 13:44:26 crc kubenswrapper[4719]: I1215 13:44:26.475702 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:44:26 crc kubenswrapper[4719]: E1215 13:44:26.476572 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:44:39 crc kubenswrapper[4719]: I1215 13:44:39.481677 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:44:39 crc kubenswrapper[4719]: E1215 13:44:39.482412 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:44:51 crc kubenswrapper[4719]: I1215 13:44:51.484196 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:44:51 crc kubenswrapper[4719]: E1215 13:44:51.485252 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.145175 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2"] Dec 15 13:45:00 crc kubenswrapper[4719]: E1215 13:45:00.145945 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerName="extract-utilities" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.145957 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerName="extract-utilities" Dec 15 13:45:00 crc kubenswrapper[4719]: E1215 13:45:00.145970 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerName="extract-content" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.145976 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerName="extract-content" Dec 15 13:45:00 crc kubenswrapper[4719]: E1215 13:45:00.146006 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerName="registry-server" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.146013 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerName="registry-server" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.146183 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a8b0687-7b15-4702-8ed7-6a52c51679d9" containerName="registry-server" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.146754 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.151248 4719 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.151372 4719 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.156919 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2"] Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.328365 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wns8f\" (UniqueName: \"kubernetes.io/projected/192a6fb7-c8da-4c83-8a46-320f9400d9b3-kube-api-access-wns8f\") pod \"collect-profiles-29430105-922t2\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.328716 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/192a6fb7-c8da-4c83-8a46-320f9400d9b3-secret-volume\") pod \"collect-profiles-29430105-922t2\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.328878 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/192a6fb7-c8da-4c83-8a46-320f9400d9b3-config-volume\") pod \"collect-profiles-29430105-922t2\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.430980 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/192a6fb7-c8da-4c83-8a46-320f9400d9b3-secret-volume\") pod \"collect-profiles-29430105-922t2\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.431080 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/192a6fb7-c8da-4c83-8a46-320f9400d9b3-config-volume\") pod \"collect-profiles-29430105-922t2\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.431166 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wns8f\" (UniqueName: \"kubernetes.io/projected/192a6fb7-c8da-4c83-8a46-320f9400d9b3-kube-api-access-wns8f\") pod \"collect-profiles-29430105-922t2\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.432034 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/192a6fb7-c8da-4c83-8a46-320f9400d9b3-config-volume\") pod \"collect-profiles-29430105-922t2\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.436607 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/192a6fb7-c8da-4c83-8a46-320f9400d9b3-secret-volume\") pod \"collect-profiles-29430105-922t2\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.452883 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wns8f\" (UniqueName: \"kubernetes.io/projected/192a6fb7-c8da-4c83-8a46-320f9400d9b3-kube-api-access-wns8f\") pod \"collect-profiles-29430105-922t2\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.471376 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:00 crc kubenswrapper[4719]: I1215 13:45:00.926106 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2"] Dec 15 13:45:00 crc kubenswrapper[4719]: W1215 13:45:00.938369 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod192a6fb7_c8da_4c83_8a46_320f9400d9b3.slice/crio-b66313d9c3c0bf1d36b03f285a8c59f4c7b8adad4b7dc4b1d44a96fac2cd55d6 WatchSource:0}: Error finding container b66313d9c3c0bf1d36b03f285a8c59f4c7b8adad4b7dc4b1d44a96fac2cd55d6: Status 404 returned error can't find the container with id b66313d9c3c0bf1d36b03f285a8c59f4c7b8adad4b7dc4b1d44a96fac2cd55d6 Dec 15 13:45:01 crc kubenswrapper[4719]: I1215 13:45:01.560534 4719 generic.go:334] "Generic (PLEG): container finished" podID="932e9f4f-4b8c-4403-aa58-feb7af89e510" containerID="e6f4b8fe43b6aa2fe7598b8bfe4e3bd03ce42ecefeca721c81a3debc70eb43fb" exitCode=0 Dec 15 13:45:01 crc kubenswrapper[4719]: I1215 13:45:01.560625 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" event={"ID":"932e9f4f-4b8c-4403-aa58-feb7af89e510","Type":"ContainerDied","Data":"e6f4b8fe43b6aa2fe7598b8bfe4e3bd03ce42ecefeca721c81a3debc70eb43fb"} Dec 15 13:45:01 crc kubenswrapper[4719]: I1215 13:45:01.565525 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" event={"ID":"192a6fb7-c8da-4c83-8a46-320f9400d9b3","Type":"ContainerDied","Data":"ca47b2fb5ffe2fdc79b36c57466e92e1f04bec96d546d4acda72f89b6ae514d7"} Dec 15 13:45:01 crc kubenswrapper[4719]: I1215 13:45:01.565576 4719 generic.go:334] "Generic (PLEG): container finished" podID="192a6fb7-c8da-4c83-8a46-320f9400d9b3" containerID="ca47b2fb5ffe2fdc79b36c57466e92e1f04bec96d546d4acda72f89b6ae514d7" exitCode=0 Dec 15 13:45:01 crc kubenswrapper[4719]: I1215 13:45:01.565625 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" event={"ID":"192a6fb7-c8da-4c83-8a46-320f9400d9b3","Type":"ContainerStarted","Data":"b66313d9c3c0bf1d36b03f285a8c59f4c7b8adad4b7dc4b1d44a96fac2cd55d6"} Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.475823 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:45:02 crc kubenswrapper[4719]: E1215 13:45:02.476314 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.685427 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.732093 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bf8hx/crc-debug-xjw9p"] Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.744341 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bf8hx/crc-debug-xjw9p"] Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.873485 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/932e9f4f-4b8c-4403-aa58-feb7af89e510-host\") pod \"932e9f4f-4b8c-4403-aa58-feb7af89e510\" (UID: \"932e9f4f-4b8c-4403-aa58-feb7af89e510\") " Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.873560 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hxv5\" (UniqueName: \"kubernetes.io/projected/932e9f4f-4b8c-4403-aa58-feb7af89e510-kube-api-access-4hxv5\") pod \"932e9f4f-4b8c-4403-aa58-feb7af89e510\" (UID: \"932e9f4f-4b8c-4403-aa58-feb7af89e510\") " Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.873765 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/932e9f4f-4b8c-4403-aa58-feb7af89e510-host" (OuterVolumeSpecName: "host") pod "932e9f4f-4b8c-4403-aa58-feb7af89e510" (UID: "932e9f4f-4b8c-4403-aa58-feb7af89e510"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.874344 4719 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/932e9f4f-4b8c-4403-aa58-feb7af89e510-host\") on node \"crc\" DevicePath \"\"" Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.880049 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/932e9f4f-4b8c-4403-aa58-feb7af89e510-kube-api-access-4hxv5" (OuterVolumeSpecName: "kube-api-access-4hxv5") pod "932e9f4f-4b8c-4403-aa58-feb7af89e510" (UID: "932e9f4f-4b8c-4403-aa58-feb7af89e510"). InnerVolumeSpecName "kube-api-access-4hxv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.974702 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.976925 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wns8f\" (UniqueName: \"kubernetes.io/projected/192a6fb7-c8da-4c83-8a46-320f9400d9b3-kube-api-access-wns8f\") pod \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.977362 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hxv5\" (UniqueName: \"kubernetes.io/projected/932e9f4f-4b8c-4403-aa58-feb7af89e510-kube-api-access-4hxv5\") on node \"crc\" DevicePath \"\"" Dec 15 13:45:02 crc kubenswrapper[4719]: I1215 13:45:02.981830 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/192a6fb7-c8da-4c83-8a46-320f9400d9b3-kube-api-access-wns8f" (OuterVolumeSpecName: "kube-api-access-wns8f") pod "192a6fb7-c8da-4c83-8a46-320f9400d9b3" (UID: "192a6fb7-c8da-4c83-8a46-320f9400d9b3"). InnerVolumeSpecName "kube-api-access-wns8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.078120 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/192a6fb7-c8da-4c83-8a46-320f9400d9b3-config-volume\") pod \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.078220 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/192a6fb7-c8da-4c83-8a46-320f9400d9b3-secret-volume\") pod \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\" (UID: \"192a6fb7-c8da-4c83-8a46-320f9400d9b3\") " Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.078575 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wns8f\" (UniqueName: \"kubernetes.io/projected/192a6fb7-c8da-4c83-8a46-320f9400d9b3-kube-api-access-wns8f\") on node \"crc\" DevicePath \"\"" Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.079143 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/192a6fb7-c8da-4c83-8a46-320f9400d9b3-config-volume" (OuterVolumeSpecName: "config-volume") pod "192a6fb7-c8da-4c83-8a46-320f9400d9b3" (UID: "192a6fb7-c8da-4c83-8a46-320f9400d9b3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.085223 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/192a6fb7-c8da-4c83-8a46-320f9400d9b3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "192a6fb7-c8da-4c83-8a46-320f9400d9b3" (UID: "192a6fb7-c8da-4c83-8a46-320f9400d9b3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.180765 4719 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/192a6fb7-c8da-4c83-8a46-320f9400d9b3-config-volume\") on node \"crc\" DevicePath \"\"" Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.180821 4719 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/192a6fb7-c8da-4c83-8a46-320f9400d9b3-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.489980 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="932e9f4f-4b8c-4403-aa58-feb7af89e510" path="/var/lib/kubelet/pods/932e9f4f-4b8c-4403-aa58-feb7af89e510/volumes" Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.583476 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-xjw9p" Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.583475 4719 scope.go:117] "RemoveContainer" containerID="e6f4b8fe43b6aa2fe7598b8bfe4e3bd03ce42ecefeca721c81a3debc70eb43fb" Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.589070 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" event={"ID":"192a6fb7-c8da-4c83-8a46-320f9400d9b3","Type":"ContainerDied","Data":"b66313d9c3c0bf1d36b03f285a8c59f4c7b8adad4b7dc4b1d44a96fac2cd55d6"} Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.589105 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b66313d9c3c0bf1d36b03f285a8c59f4c7b8adad4b7dc4b1d44a96fac2cd55d6" Dec 15 13:45:03 crc kubenswrapper[4719]: I1215 13:45:03.589161 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29430105-922t2" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.074561 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29"] Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.087215 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29430060-l5g29"] Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.096409 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bf8hx/crc-debug-5dz29"] Dec 15 13:45:04 crc kubenswrapper[4719]: E1215 13:45:04.096799 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="932e9f4f-4b8c-4403-aa58-feb7af89e510" containerName="container-00" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.096817 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="932e9f4f-4b8c-4403-aa58-feb7af89e510" containerName="container-00" Dec 15 13:45:04 crc kubenswrapper[4719]: E1215 13:45:04.096871 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="192a6fb7-c8da-4c83-8a46-320f9400d9b3" containerName="collect-profiles" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.096879 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="192a6fb7-c8da-4c83-8a46-320f9400d9b3" containerName="collect-profiles" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.097059 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="932e9f4f-4b8c-4403-aa58-feb7af89e510" containerName="container-00" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.097092 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="192a6fb7-c8da-4c83-8a46-320f9400d9b3" containerName="collect-profiles" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.097683 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-5dz29" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.195538 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwmb2\" (UniqueName: \"kubernetes.io/projected/b9104627-7b6e-4155-8492-55a7c595f58a-kube-api-access-xwmb2\") pod \"crc-debug-5dz29\" (UID: \"b9104627-7b6e-4155-8492-55a7c595f58a\") " pod="openshift-must-gather-bf8hx/crc-debug-5dz29" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.195674 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9104627-7b6e-4155-8492-55a7c595f58a-host\") pod \"crc-debug-5dz29\" (UID: \"b9104627-7b6e-4155-8492-55a7c595f58a\") " pod="openshift-must-gather-bf8hx/crc-debug-5dz29" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.297049 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwmb2\" (UniqueName: \"kubernetes.io/projected/b9104627-7b6e-4155-8492-55a7c595f58a-kube-api-access-xwmb2\") pod \"crc-debug-5dz29\" (UID: \"b9104627-7b6e-4155-8492-55a7c595f58a\") " pod="openshift-must-gather-bf8hx/crc-debug-5dz29" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.297181 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9104627-7b6e-4155-8492-55a7c595f58a-host\") pod \"crc-debug-5dz29\" (UID: \"b9104627-7b6e-4155-8492-55a7c595f58a\") " pod="openshift-must-gather-bf8hx/crc-debug-5dz29" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.297291 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9104627-7b6e-4155-8492-55a7c595f58a-host\") pod \"crc-debug-5dz29\" (UID: \"b9104627-7b6e-4155-8492-55a7c595f58a\") " pod="openshift-must-gather-bf8hx/crc-debug-5dz29" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.316384 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwmb2\" (UniqueName: \"kubernetes.io/projected/b9104627-7b6e-4155-8492-55a7c595f58a-kube-api-access-xwmb2\") pod \"crc-debug-5dz29\" (UID: \"b9104627-7b6e-4155-8492-55a7c595f58a\") " pod="openshift-must-gather-bf8hx/crc-debug-5dz29" Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.439923 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-5dz29" Dec 15 13:45:04 crc kubenswrapper[4719]: W1215 13:45:04.468988 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9104627_7b6e_4155_8492_55a7c595f58a.slice/crio-f362ca38535ddb4900ab14b566b5f4e1e2f8a63e7e7d8a1e65a050fa60313805 WatchSource:0}: Error finding container f362ca38535ddb4900ab14b566b5f4e1e2f8a63e7e7d8a1e65a050fa60313805: Status 404 returned error can't find the container with id f362ca38535ddb4900ab14b566b5f4e1e2f8a63e7e7d8a1e65a050fa60313805 Dec 15 13:45:04 crc kubenswrapper[4719]: I1215 13:45:04.622406 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/crc-debug-5dz29" event={"ID":"b9104627-7b6e-4155-8492-55a7c595f58a","Type":"ContainerStarted","Data":"f362ca38535ddb4900ab14b566b5f4e1e2f8a63e7e7d8a1e65a050fa60313805"} Dec 15 13:45:05 crc kubenswrapper[4719]: I1215 13:45:05.486733 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29f0f318-9ef6-4f33-aec9-10cbaee5d0e7" path="/var/lib/kubelet/pods/29f0f318-9ef6-4f33-aec9-10cbaee5d0e7/volumes" Dec 15 13:45:05 crc kubenswrapper[4719]: I1215 13:45:05.631306 4719 generic.go:334] "Generic (PLEG): container finished" podID="b9104627-7b6e-4155-8492-55a7c595f58a" containerID="b7025df69cbb1837897f9c27b60cdcd8ea3628e1e45d299c0d124e7ce41f9a5f" exitCode=0 Dec 15 13:45:05 crc kubenswrapper[4719]: I1215 13:45:05.631398 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/crc-debug-5dz29" event={"ID":"b9104627-7b6e-4155-8492-55a7c595f58a","Type":"ContainerDied","Data":"b7025df69cbb1837897f9c27b60cdcd8ea3628e1e45d299c0d124e7ce41f9a5f"} Dec 15 13:45:06 crc kubenswrapper[4719]: I1215 13:45:06.748480 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-5dz29" Dec 15 13:45:06 crc kubenswrapper[4719]: I1215 13:45:06.939247 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9104627-7b6e-4155-8492-55a7c595f58a-host\") pod \"b9104627-7b6e-4155-8492-55a7c595f58a\" (UID: \"b9104627-7b6e-4155-8492-55a7c595f58a\") " Dec 15 13:45:06 crc kubenswrapper[4719]: I1215 13:45:06.939876 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9104627-7b6e-4155-8492-55a7c595f58a-host" (OuterVolumeSpecName: "host") pod "b9104627-7b6e-4155-8492-55a7c595f58a" (UID: "b9104627-7b6e-4155-8492-55a7c595f58a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 13:45:06 crc kubenswrapper[4719]: I1215 13:45:06.940992 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwmb2\" (UniqueName: \"kubernetes.io/projected/b9104627-7b6e-4155-8492-55a7c595f58a-kube-api-access-xwmb2\") pod \"b9104627-7b6e-4155-8492-55a7c595f58a\" (UID: \"b9104627-7b6e-4155-8492-55a7c595f58a\") " Dec 15 13:45:06 crc kubenswrapper[4719]: I1215 13:45:06.941708 4719 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9104627-7b6e-4155-8492-55a7c595f58a-host\") on node \"crc\" DevicePath \"\"" Dec 15 13:45:06 crc kubenswrapper[4719]: I1215 13:45:06.959132 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9104627-7b6e-4155-8492-55a7c595f58a-kube-api-access-xwmb2" (OuterVolumeSpecName: "kube-api-access-xwmb2") pod "b9104627-7b6e-4155-8492-55a7c595f58a" (UID: "b9104627-7b6e-4155-8492-55a7c595f58a"). InnerVolumeSpecName "kube-api-access-xwmb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:45:07 crc kubenswrapper[4719]: I1215 13:45:07.046253 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwmb2\" (UniqueName: \"kubernetes.io/projected/b9104627-7b6e-4155-8492-55a7c595f58a-kube-api-access-xwmb2\") on node \"crc\" DevicePath \"\"" Dec 15 13:45:07 crc kubenswrapper[4719]: I1215 13:45:07.623786 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bf8hx/crc-debug-5dz29"] Dec 15 13:45:07 crc kubenswrapper[4719]: I1215 13:45:07.637709 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bf8hx/crc-debug-5dz29"] Dec 15 13:45:07 crc kubenswrapper[4719]: I1215 13:45:07.660205 4719 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f362ca38535ddb4900ab14b566b5f4e1e2f8a63e7e7d8a1e65a050fa60313805" Dec 15 13:45:07 crc kubenswrapper[4719]: I1215 13:45:07.660280 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-5dz29" Dec 15 13:45:08 crc kubenswrapper[4719]: I1215 13:45:08.869405 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bf8hx/crc-debug-bhjvh"] Dec 15 13:45:08 crc kubenswrapper[4719]: E1215 13:45:08.869806 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9104627-7b6e-4155-8492-55a7c595f58a" containerName="container-00" Dec 15 13:45:08 crc kubenswrapper[4719]: I1215 13:45:08.869818 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9104627-7b6e-4155-8492-55a7c595f58a" containerName="container-00" Dec 15 13:45:08 crc kubenswrapper[4719]: I1215 13:45:08.870018 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9104627-7b6e-4155-8492-55a7c595f58a" containerName="container-00" Dec 15 13:45:08 crc kubenswrapper[4719]: I1215 13:45:08.870587 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" Dec 15 13:45:08 crc kubenswrapper[4719]: I1215 13:45:08.988460 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dfd55b12-b976-45f5-b07f-abcb95a350e3-host\") pod \"crc-debug-bhjvh\" (UID: \"dfd55b12-b976-45f5-b07f-abcb95a350e3\") " pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" Dec 15 13:45:08 crc kubenswrapper[4719]: I1215 13:45:08.988569 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s86sl\" (UniqueName: \"kubernetes.io/projected/dfd55b12-b976-45f5-b07f-abcb95a350e3-kube-api-access-s86sl\") pod \"crc-debug-bhjvh\" (UID: \"dfd55b12-b976-45f5-b07f-abcb95a350e3\") " pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" Dec 15 13:45:09 crc kubenswrapper[4719]: I1215 13:45:09.108549 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dfd55b12-b976-45f5-b07f-abcb95a350e3-host\") pod \"crc-debug-bhjvh\" (UID: \"dfd55b12-b976-45f5-b07f-abcb95a350e3\") " pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" Dec 15 13:45:09 crc kubenswrapper[4719]: I1215 13:45:09.108693 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dfd55b12-b976-45f5-b07f-abcb95a350e3-host\") pod \"crc-debug-bhjvh\" (UID: \"dfd55b12-b976-45f5-b07f-abcb95a350e3\") " pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" Dec 15 13:45:09 crc kubenswrapper[4719]: I1215 13:45:09.108827 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s86sl\" (UniqueName: \"kubernetes.io/projected/dfd55b12-b976-45f5-b07f-abcb95a350e3-kube-api-access-s86sl\") pod \"crc-debug-bhjvh\" (UID: \"dfd55b12-b976-45f5-b07f-abcb95a350e3\") " pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" Dec 15 13:45:09 crc kubenswrapper[4719]: I1215 13:45:09.128063 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s86sl\" (UniqueName: \"kubernetes.io/projected/dfd55b12-b976-45f5-b07f-abcb95a350e3-kube-api-access-s86sl\") pod \"crc-debug-bhjvh\" (UID: \"dfd55b12-b976-45f5-b07f-abcb95a350e3\") " pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" Dec 15 13:45:09 crc kubenswrapper[4719]: I1215 13:45:09.186465 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" Dec 15 13:45:09 crc kubenswrapper[4719]: W1215 13:45:09.242720 4719 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfd55b12_b976_45f5_b07f_abcb95a350e3.slice/crio-38b56e1707e15c235721c7b194e05dabc2ff11dec1c60932812c9cffae725386 WatchSource:0}: Error finding container 38b56e1707e15c235721c7b194e05dabc2ff11dec1c60932812c9cffae725386: Status 404 returned error can't find the container with id 38b56e1707e15c235721c7b194e05dabc2ff11dec1c60932812c9cffae725386 Dec 15 13:45:09 crc kubenswrapper[4719]: I1215 13:45:09.490603 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9104627-7b6e-4155-8492-55a7c595f58a" path="/var/lib/kubelet/pods/b9104627-7b6e-4155-8492-55a7c595f58a/volumes" Dec 15 13:45:09 crc kubenswrapper[4719]: I1215 13:45:09.678465 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" event={"ID":"dfd55b12-b976-45f5-b07f-abcb95a350e3","Type":"ContainerStarted","Data":"38b56e1707e15c235721c7b194e05dabc2ff11dec1c60932812c9cffae725386"} Dec 15 13:45:10 crc kubenswrapper[4719]: I1215 13:45:10.687951 4719 generic.go:334] "Generic (PLEG): container finished" podID="dfd55b12-b976-45f5-b07f-abcb95a350e3" containerID="dbd871523173674be85b0bc1b682ec092c294ed065e3517be37577318caaa32a" exitCode=0 Dec 15 13:45:10 crc kubenswrapper[4719]: I1215 13:45:10.688049 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" event={"ID":"dfd55b12-b976-45f5-b07f-abcb95a350e3","Type":"ContainerDied","Data":"dbd871523173674be85b0bc1b682ec092c294ed065e3517be37577318caaa32a"} Dec 15 13:45:10 crc kubenswrapper[4719]: I1215 13:45:10.730439 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bf8hx/crc-debug-bhjvh"] Dec 15 13:45:10 crc kubenswrapper[4719]: I1215 13:45:10.738928 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bf8hx/crc-debug-bhjvh"] Dec 15 13:45:11 crc kubenswrapper[4719]: I1215 13:45:11.795221 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" Dec 15 13:45:11 crc kubenswrapper[4719]: I1215 13:45:11.963472 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dfd55b12-b976-45f5-b07f-abcb95a350e3-host\") pod \"dfd55b12-b976-45f5-b07f-abcb95a350e3\" (UID: \"dfd55b12-b976-45f5-b07f-abcb95a350e3\") " Dec 15 13:45:11 crc kubenswrapper[4719]: I1215 13:45:11.963700 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s86sl\" (UniqueName: \"kubernetes.io/projected/dfd55b12-b976-45f5-b07f-abcb95a350e3-kube-api-access-s86sl\") pod \"dfd55b12-b976-45f5-b07f-abcb95a350e3\" (UID: \"dfd55b12-b976-45f5-b07f-abcb95a350e3\") " Dec 15 13:45:11 crc kubenswrapper[4719]: I1215 13:45:11.964529 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dfd55b12-b976-45f5-b07f-abcb95a350e3-host" (OuterVolumeSpecName: "host") pod "dfd55b12-b976-45f5-b07f-abcb95a350e3" (UID: "dfd55b12-b976-45f5-b07f-abcb95a350e3"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 15 13:45:11 crc kubenswrapper[4719]: I1215 13:45:11.977230 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfd55b12-b976-45f5-b07f-abcb95a350e3-kube-api-access-s86sl" (OuterVolumeSpecName: "kube-api-access-s86sl") pod "dfd55b12-b976-45f5-b07f-abcb95a350e3" (UID: "dfd55b12-b976-45f5-b07f-abcb95a350e3"). InnerVolumeSpecName "kube-api-access-s86sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:45:12 crc kubenswrapper[4719]: I1215 13:45:12.066826 4719 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dfd55b12-b976-45f5-b07f-abcb95a350e3-host\") on node \"crc\" DevicePath \"\"" Dec 15 13:45:12 crc kubenswrapper[4719]: I1215 13:45:12.067074 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s86sl\" (UniqueName: \"kubernetes.io/projected/dfd55b12-b976-45f5-b07f-abcb95a350e3-kube-api-access-s86sl\") on node \"crc\" DevicePath \"\"" Dec 15 13:45:12 crc kubenswrapper[4719]: I1215 13:45:12.705610 4719 scope.go:117] "RemoveContainer" containerID="dbd871523173674be85b0bc1b682ec092c294ed065e3517be37577318caaa32a" Dec 15 13:45:12 crc kubenswrapper[4719]: I1215 13:45:12.705629 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/crc-debug-bhjvh" Dec 15 13:45:13 crc kubenswrapper[4719]: I1215 13:45:13.475818 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:45:13 crc kubenswrapper[4719]: E1215 13:45:13.476404 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:45:13 crc kubenswrapper[4719]: I1215 13:45:13.494108 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfd55b12-b976-45f5-b07f-abcb95a350e3" path="/var/lib/kubelet/pods/dfd55b12-b976-45f5-b07f-abcb95a350e3/volumes" Dec 15 13:45:27 crc kubenswrapper[4719]: I1215 13:45:27.476507 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:45:27 crc kubenswrapper[4719]: E1215 13:45:27.477276 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:45:36 crc kubenswrapper[4719]: I1215 13:45:36.419570 4719 scope.go:117] "RemoveContainer" containerID="8a6c6ec5b74c2ca53223dcc984ead8444d631af836779f9d69a97ad64d648c11" Dec 15 13:45:39 crc kubenswrapper[4719]: I1215 13:45:39.482067 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:45:39 crc kubenswrapper[4719]: E1215 13:45:39.482874 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:45:43 crc kubenswrapper[4719]: I1215 13:45:43.504002 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7ff7556d66-bpsdl_a996feea-1c4e-496f-a91b-2f07b852d22e/barbican-api/0.log" Dec 15 13:45:43 crc kubenswrapper[4719]: I1215 13:45:43.660845 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7ff7556d66-bpsdl_a996feea-1c4e-496f-a91b-2f07b852d22e/barbican-api-log/0.log" Dec 15 13:45:43 crc kubenswrapper[4719]: I1215 13:45:43.824295 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b87454b66-vfv5x_2da95a03-4fe4-4674-b56c-f697d4ab3045/barbican-keystone-listener/0.log" Dec 15 13:45:43 crc kubenswrapper[4719]: I1215 13:45:43.862270 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b87454b66-vfv5x_2da95a03-4fe4-4674-b56c-f697d4ab3045/barbican-keystone-listener-log/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.009915 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-769f77d87-tfk84_d2f1a60c-ffe6-4563-9466-2a3129be55ee/barbican-worker/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.112284 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-769f77d87-tfk84_d2f1a60c-ffe6-4563-9466-2a3129be55ee/barbican-worker-log/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.219652 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4d48w_19e3b1bb-a052-47fd-a833-bed95ef223fc/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.349911 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0/ceilometer-central-agent/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.464629 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0/proxy-httpd/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.500783 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0/ceilometer-notification-agent/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.583719 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_25f7ade1-fa78-4352-9f9c-93a2e0e5b0c0/sg-core/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.695482 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6133c925-0cd7-489d-8b32-097e9b78ad70/cinder-api-log/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.817675 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6133c925-0cd7-489d-8b32-097e9b78ad70/cinder-api/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.978020 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d60e91f3-5bb1-4958-8489-e047f9a31c87/cinder-scheduler/0.log" Dec 15 13:45:44 crc kubenswrapper[4719]: I1215 13:45:44.990509 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d60e91f3-5bb1-4958-8489-e047f9a31c87/probe/0.log" Dec 15 13:45:45 crc kubenswrapper[4719]: I1215 13:45:45.167187 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-7q45z_2b3c8271-0d1c-4da0-b282-824968e884bb/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:45 crc kubenswrapper[4719]: I1215 13:45:45.223818 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kqrtj_611b65c2-c554-4f37-a644-41bcbd27ad46/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:45 crc kubenswrapper[4719]: I1215 13:45:45.394493 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-cvlz2_3cd29959-bbda-441f-901a-726fc4ebb4c4/init/0.log" Dec 15 13:45:45 crc kubenswrapper[4719]: I1215 13:45:45.695578 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-zn8ct_b72d7eb2-0c87-4377-8338-118b53d43241/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:45 crc kubenswrapper[4719]: I1215 13:45:45.764022 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-cvlz2_3cd29959-bbda-441f-901a-726fc4ebb4c4/init/0.log" Dec 15 13:45:45 crc kubenswrapper[4719]: I1215 13:45:45.827994 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-cvlz2_3cd29959-bbda-441f-901a-726fc4ebb4c4/dnsmasq-dns/0.log" Dec 15 13:45:46 crc kubenswrapper[4719]: I1215 13:45:46.087794 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fc2e0fce-d6e0-47e6-bd88-34c45829436a/glance-httpd/0.log" Dec 15 13:45:46 crc kubenswrapper[4719]: I1215 13:45:46.095446 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fc2e0fce-d6e0-47e6-bd88-34c45829436a/glance-log/0.log" Dec 15 13:45:46 crc kubenswrapper[4719]: I1215 13:45:46.322235 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f26a2026-ce1a-4411-8e19-e1981be6f927/glance-httpd/0.log" Dec 15 13:45:46 crc kubenswrapper[4719]: I1215 13:45:46.464226 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f26a2026-ce1a-4411-8e19-e1981be6f927/glance-log/0.log" Dec 15 13:45:46 crc kubenswrapper[4719]: I1215 13:45:46.533156 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6bc5fb48-jqf6b_f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926/horizon/2.log" Dec 15 13:45:46 crc kubenswrapper[4719]: I1215 13:45:46.683967 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6bc5fb48-jqf6b_f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926/horizon/1.log" Dec 15 13:45:46 crc kubenswrapper[4719]: I1215 13:45:46.887782 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-v9x9s_71b47db7-3079-4990-8a37-b02ea6f0368b/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:47 crc kubenswrapper[4719]: I1215 13:45:47.189901 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6bc5fb48-jqf6b_f9b7dc15-da8d-41ad-a5ab-cb7e16b0c926/horizon-log/0.log" Dec 15 13:45:47 crc kubenswrapper[4719]: I1215 13:45:47.308220 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-gv6zg_2e7c3312-b561-431c-8794-c63098dbe33d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:47 crc kubenswrapper[4719]: I1215 13:45:47.653926 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29430061-9mfp2_32ba5a78-e7da-4c5c-9cdc-cf3e4eb18992/keystone-cron/0.log" Dec 15 13:45:47 crc kubenswrapper[4719]: I1215 13:45:47.831981 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5c7789bfff-vvvqf_e087fc68-a428-4d63-b1c1-9d9a6b7b0c93/keystone-api/0.log" Dec 15 13:45:47 crc kubenswrapper[4719]: I1215 13:45:47.869266 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_9a236191-195f-4941-b740-ac7ee096858a/kube-state-metrics/0.log" Dec 15 13:45:48 crc kubenswrapper[4719]: I1215 13:45:48.061456 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-6j7ql_c390d0c5-24b7-4810-baf1-d7b9fd9dd13f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:48 crc kubenswrapper[4719]: I1215 13:45:48.864103 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5668b96c97-48pkv_4ddfe7f4-2192-4520-bfae-028a2c41e459/neutron-httpd/0.log" Dec 15 13:45:48 crc kubenswrapper[4719]: I1215 13:45:48.882188 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-9pxkg_39093fe3-9551-4d58-9cd0-607f1694a71c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:49 crc kubenswrapper[4719]: I1215 13:45:49.280680 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5668b96c97-48pkv_4ddfe7f4-2192-4520-bfae-028a2c41e459/neutron-api/0.log" Dec 15 13:45:49 crc kubenswrapper[4719]: I1215 13:45:49.981956 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d2da200b-26c0-4242-a443-fa579e6da8da/nova-cell0-conductor-conductor/0.log" Dec 15 13:45:50 crc kubenswrapper[4719]: I1215 13:45:50.115462 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_36c7f0b4-980a-4543-a7bc-38e6115a9d53/nova-cell1-conductor-conductor/0.log" Dec 15 13:45:50 crc kubenswrapper[4719]: I1215 13:45:50.551202 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ccc69b5f-a98b-4188-b08f-4cb6f58c085c/nova-cell1-novncproxy-novncproxy/0.log" Dec 15 13:45:50 crc kubenswrapper[4719]: I1215 13:45:50.577372 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_59d59387-003d-47ae-a322-48787db59367/nova-api-log/0.log" Dec 15 13:45:50 crc kubenswrapper[4719]: I1215 13:45:50.906763 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-hvqz9_48c18077-569a-4bbd-ace2-64dbf06bcbee/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:51 crc kubenswrapper[4719]: I1215 13:45:51.095179 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_59d59387-003d-47ae-a322-48787db59367/nova-api-api/0.log" Dec 15 13:45:51 crc kubenswrapper[4719]: I1215 13:45:51.136915 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_45c8d521-b7f2-4f93-a6e8-94d6aaf85390/nova-metadata-log/0.log" Dec 15 13:45:51 crc kubenswrapper[4719]: I1215 13:45:51.480878 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:45:51 crc kubenswrapper[4719]: E1215 13:45:51.481272 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:45:51 crc kubenswrapper[4719]: I1215 13:45:51.846689 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9655f31a-6eae-461b-932a-0b1b4ca56048/mysql-bootstrap/0.log" Dec 15 13:45:52 crc kubenswrapper[4719]: I1215 13:45:52.139290 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9655f31a-6eae-461b-932a-0b1b4ca56048/galera/0.log" Dec 15 13:45:52 crc kubenswrapper[4719]: I1215 13:45:52.151841 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_0945867e-d6c4-4f3e-976d-b2ae68497a11/nova-scheduler-scheduler/0.log" Dec 15 13:45:52 crc kubenswrapper[4719]: I1215 13:45:52.157280 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9655f31a-6eae-461b-932a-0b1b4ca56048/mysql-bootstrap/0.log" Dec 15 13:45:52 crc kubenswrapper[4719]: I1215 13:45:52.475315 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b614221-314c-4033-80d8-20738c1b92ae/mysql-bootstrap/0.log" Dec 15 13:45:52 crc kubenswrapper[4719]: I1215 13:45:52.730536 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b614221-314c-4033-80d8-20738c1b92ae/mysql-bootstrap/0.log" Dec 15 13:45:52 crc kubenswrapper[4719]: I1215 13:45:52.766169 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b614221-314c-4033-80d8-20738c1b92ae/galera/0.log" Dec 15 13:45:53 crc kubenswrapper[4719]: I1215 13:45:53.003373 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0b238728-d6aa-4b27-b48b-beb0cc7ff437/openstackclient/0.log" Dec 15 13:45:53 crc kubenswrapper[4719]: I1215 13:45:53.171920 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4ht64_4c7986d1-97c7-4b8c-8b2f-594fafc24991/ovn-controller/0.log" Dec 15 13:45:53 crc kubenswrapper[4719]: I1215 13:45:53.405492 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-84b56_8dc6d84f-343b-456a-bad8-3171dccc2b9d/openstack-network-exporter/0.log" Dec 15 13:45:53 crc kubenswrapper[4719]: I1215 13:45:53.560672 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_45c8d521-b7f2-4f93-a6e8-94d6aaf85390/nova-metadata-metadata/0.log" Dec 15 13:45:53 crc kubenswrapper[4719]: I1215 13:45:53.726731 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hrrfr_d60bc22b-09f1-4b84-84ad-a9ee31a18a43/ovsdb-server-init/0.log" Dec 15 13:45:53 crc kubenswrapper[4719]: I1215 13:45:53.928700 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hrrfr_d60bc22b-09f1-4b84-84ad-a9ee31a18a43/ovsdb-server-init/0.log" Dec 15 13:45:53 crc kubenswrapper[4719]: I1215 13:45:53.998550 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hrrfr_d60bc22b-09f1-4b84-84ad-a9ee31a18a43/ovs-vswitchd/0.log" Dec 15 13:45:53 crc kubenswrapper[4719]: I1215 13:45:53.999968 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-hrrfr_d60bc22b-09f1-4b84-84ad-a9ee31a18a43/ovsdb-server/0.log" Dec 15 13:45:54 crc kubenswrapper[4719]: I1215 13:45:54.309364 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-nrfr7_87e0fa54-5d1a-4e53-aef8-257af1ebd4d3/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:54 crc kubenswrapper[4719]: I1215 13:45:54.367167 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a0b079a1-3e14-4457-aac6-213038651316/openstack-network-exporter/0.log" Dec 15 13:45:54 crc kubenswrapper[4719]: I1215 13:45:54.522969 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a0b079a1-3e14-4457-aac6-213038651316/ovn-northd/0.log" Dec 15 13:45:54 crc kubenswrapper[4719]: I1215 13:45:54.632181 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_66e0622c-a031-4927-83f3-d7d4c8dc0233/openstack-network-exporter/0.log" Dec 15 13:45:54 crc kubenswrapper[4719]: I1215 13:45:54.735047 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_66e0622c-a031-4927-83f3-d7d4c8dc0233/ovsdbserver-nb/0.log" Dec 15 13:45:54 crc kubenswrapper[4719]: I1215 13:45:54.759744 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c9dd6881-cb99-4c3f-9a05-75c9ae4f4da1/memcached/0.log" Dec 15 13:45:54 crc kubenswrapper[4719]: I1215 13:45:54.844516 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_66f93b1d-4c36-4f90-ac6c-28b302a8c34b/openstack-network-exporter/0.log" Dec 15 13:45:54 crc kubenswrapper[4719]: I1215 13:45:54.922748 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_66f93b1d-4c36-4f90-ac6c-28b302a8c34b/ovsdbserver-sb/0.log" Dec 15 13:45:55 crc kubenswrapper[4719]: I1215 13:45:55.116723 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9d9f8bcf4-hj8fd_ac288046-bbc8-4521-84b8-c16e7d096217/placement-api/0.log" Dec 15 13:45:55 crc kubenswrapper[4719]: I1215 13:45:55.217438 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82c8c133-e1d6-4b56-b745-1aed017d0bfb/setup-container/0.log" Dec 15 13:45:55 crc kubenswrapper[4719]: I1215 13:45:55.268435 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9d9f8bcf4-hj8fd_ac288046-bbc8-4521-84b8-c16e7d096217/placement-log/0.log" Dec 15 13:45:55 crc kubenswrapper[4719]: I1215 13:45:55.556260 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82c8c133-e1d6-4b56-b745-1aed017d0bfb/rabbitmq/0.log" Dec 15 13:45:55 crc kubenswrapper[4719]: I1215 13:45:55.559935 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_82c8c133-e1d6-4b56-b745-1aed017d0bfb/setup-container/0.log" Dec 15 13:45:55 crc kubenswrapper[4719]: I1215 13:45:55.560600 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d4163ed6-9550-4d10-baab-b92859fa07df/setup-container/0.log" Dec 15 13:45:55 crc kubenswrapper[4719]: I1215 13:45:55.777265 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d4163ed6-9550-4d10-baab-b92859fa07df/setup-container/0.log" Dec 15 13:45:55 crc kubenswrapper[4719]: I1215 13:45:55.802960 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d4163ed6-9550-4d10-baab-b92859fa07df/rabbitmq/0.log" Dec 15 13:45:55 crc kubenswrapper[4719]: I1215 13:45:55.817999 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-4bhsk_f1474b5e-e320-4f22-97ae-729dbdd97939/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.056548 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-xjgnt_ed08a59f-ed8b-4195-b916-0d04c3e01f9f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.123023 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-9nf7s_588db7b9-6a66-4c02-be83-8766e47c211c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.249910 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-kfrcr_57d95048-e4af-466c-84ff-f2d00d3b5f58/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.342780 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-xhxtk_c7b477ac-26a4-4ddb-af83-57d931fe0e5a/ssh-known-hosts-edpm-deployment/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.569418 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-dcbfk_fb01ec8f-cc85-4146-9677-344c14c2a545/swift-ring-rebalance/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.602481 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7f6dc9444c-2jcmv_60352f3a-8ee6-498d-9ee3-152cc4ab562f/proxy-httpd/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.607145 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7f6dc9444c-2jcmv_60352f3a-8ee6-498d-9ee3-152cc4ab562f/proxy-server/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.829267 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/account-server/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.829496 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/account-auditor/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.836744 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/account-reaper/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.909374 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/account-replicator/0.log" Dec 15 13:45:56 crc kubenswrapper[4719]: I1215 13:45:56.967520 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/container-auditor/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.065786 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/container-server/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.107334 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/container-updater/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.138536 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/container-replicator/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.218552 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/object-auditor/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.252946 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/object-expirer/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.354791 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/object-replicator/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.383021 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/object-server/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.409507 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/object-updater/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.472930 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/rsync/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.515189 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_651edea8-3300-463c-a0fa-abd75f6056e0/swift-recon-cron/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.729347 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-x4lsb_e70ef11b-7e5a-47c9-9f63-15dd69326073/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.752918 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_8b81902c-c979-4e03-a40f-cac54a63d6db/tempest-tests-tempest-tests-runner/0.log" Dec 15 13:45:57 crc kubenswrapper[4719]: I1215 13:45:57.996004 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_d347fbd6-e218-4153-be3f-f912f19123bb/test-operator-logs-container/0.log" Dec 15 13:45:58 crc kubenswrapper[4719]: I1215 13:45:58.013994 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-rz5jl_2c709af4-4a8c-4cb9-955e-37dfe843569b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 15 13:46:02 crc kubenswrapper[4719]: I1215 13:46:02.476849 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:46:02 crc kubenswrapper[4719]: E1215 13:46:02.477509 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:46:17 crc kubenswrapper[4719]: I1215 13:46:17.475682 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:46:17 crc kubenswrapper[4719]: E1215 13:46:17.476408 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:46:24 crc kubenswrapper[4719]: I1215 13:46:24.720493 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/util/0.log" Dec 15 13:46:24 crc kubenswrapper[4719]: I1215 13:46:24.985456 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/pull/0.log" Dec 15 13:46:24 crc kubenswrapper[4719]: I1215 13:46:24.993366 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/pull/0.log" Dec 15 13:46:25 crc kubenswrapper[4719]: I1215 13:46:25.026358 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/util/0.log" Dec 15 13:46:25 crc kubenswrapper[4719]: I1215 13:46:25.210145 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/extract/0.log" Dec 15 13:46:25 crc kubenswrapper[4719]: I1215 13:46:25.228285 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/util/0.log" Dec 15 13:46:25 crc kubenswrapper[4719]: I1215 13:46:25.283026 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b077d25ccf2f5210efb6719819bf7812258d0cb933a3e1bfbdd6e0f2c9xhh6t_c62f4038-b4ab-4211-8b78-e0ffe50814b7/pull/0.log" Dec 15 13:46:25 crc kubenswrapper[4719]: I1215 13:46:25.473312 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-95949466-k2fz4_5c7fa683-b59c-45ef-bc29-92400f43e4b8/manager/0.log" Dec 15 13:46:25 crc kubenswrapper[4719]: I1215 13:46:25.563808 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5cf45c46bd-pmmwk_7f47c007-e5bf-47dd-8905-69ed7cfc3a9a/manager/0.log" Dec 15 13:46:25 crc kubenswrapper[4719]: I1215 13:46:25.706398 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66f8b87655-5td8v_7ced32f3-ca3c-406c-966b-93be87cd6d25/manager/0.log" Dec 15 13:46:25 crc kubenswrapper[4719]: I1215 13:46:25.869581 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-767f9d7567-2866h_345b77d5-0419-433e-b7d4-053a0b6fcf86/manager/0.log" Dec 15 13:46:26 crc kubenswrapper[4719]: I1215 13:46:26.051665 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-59b8dcb766-llbls_6cdd3328-cd4d-4e90-8ab9-d0953cab73de/manager/0.log" Dec 15 13:46:26 crc kubenswrapper[4719]: I1215 13:46:26.134774 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6ccf486b9-scntm_15643ba0-8573-4f51-8ee6-9fd78e10b6a0/manager/0.log" Dec 15 13:46:26 crc kubenswrapper[4719]: I1215 13:46:26.490404 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-f458558d7-zkv6s_e6adc1e0-5c73-431d-a94d-e2eabc16dd03/manager/0.log" Dec 15 13:46:26 crc kubenswrapper[4719]: I1215 13:46:26.573428 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-58944d7758-wh4dh_b6efe10d-9999-4340-822b-48fb8fb99f14/manager/0.log" Dec 15 13:46:26 crc kubenswrapper[4719]: I1215 13:46:26.801698 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5c7cbf548f-5slz5_9df14cd5-dea3-481b-9a52-1eeb3d1d0fbc/manager/0.log" Dec 15 13:46:26 crc kubenswrapper[4719]: I1215 13:46:26.843486 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5fdd9786f7-89zrh_571dd62a-58e8-4dab-ad04-a95621a65078/manager/0.log" Dec 15 13:46:27 crc kubenswrapper[4719]: I1215 13:46:27.287250 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f76f4954c-zhvw6_def1648e-ccf3-4a95-9a44-92bf0f83cb49/manager/0.log" Dec 15 13:46:27 crc kubenswrapper[4719]: I1215 13:46:27.483619 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7cd87b778f-hb66s_83a0cff1-8c43-4bb1-aa4f-66c4662fd235/manager/0.log" Dec 15 13:46:27 crc kubenswrapper[4719]: I1215 13:46:27.721026 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5fbbf8b6cc-c9szj_828d7a4d-0956-40fc-967f-e52d1701953c/manager/0.log" Dec 15 13:46:27 crc kubenswrapper[4719]: I1215 13:46:27.778454 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-68c649d9d-p7z6f_3b66851a-75d1-46ea-8d06-ad7385fdff6b/manager/0.log" Dec 15 13:46:27 crc kubenswrapper[4719]: I1215 13:46:27.944693 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7b67c7f6c5cz556_24c5b173-c013-45e1-9563-1e49556b23a5/manager/0.log" Dec 15 13:46:28 crc kubenswrapper[4719]: I1215 13:46:28.417209 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-66c467f856-n9b75_131429f3-e15e-4107-af0e-b2d4c01f87af/operator/0.log" Dec 15 13:46:28 crc kubenswrapper[4719]: I1215 13:46:28.566496 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zc4xm_bcaba66e-19b2-4c6f-9ec2-e545a7531624/registry-server/0.log" Dec 15 13:46:29 crc kubenswrapper[4719]: I1215 13:46:29.014971 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bf6d4f946-gn5n9_d0a679bd-fdb2-4a1f-97ae-8a31945f4c2e/manager/0.log" Dec 15 13:46:29 crc kubenswrapper[4719]: I1215 13:46:29.183652 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8665b56d78-wslbg_9adcd81c-e72b-44b8-bd4f-36e49c8e5c82/manager/0.log" Dec 15 13:46:29 crc kubenswrapper[4719]: I1215 13:46:29.206594 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7b4b554f64-x55hj_db19bab3-733c-4ef1-8f46-df4cf5842701/manager/0.log" Dec 15 13:46:29 crc kubenswrapper[4719]: I1215 13:46:29.326680 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-jzg2t_0c36396e-131f-4154-b82e-490b36d7fa63/operator/0.log" Dec 15 13:46:29 crc kubenswrapper[4719]: I1215 13:46:29.524241 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5c6df8f9-d9d67_c42cf0cb-4f9f-4c6a-ad07-ed33b7b13157/manager/0.log" Dec 15 13:46:29 crc kubenswrapper[4719]: I1215 13:46:29.672082 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-97d456b9-qxhfl_375826ba-0785-4935-96c2-a2e5345aaf45/manager/0.log" Dec 15 13:46:29 crc kubenswrapper[4719]: I1215 13:46:29.781786 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-756ccf86c7-l98v4_72cd4bb7-5783-492b-bae4-1a9fa633cd7b/manager/0.log" Dec 15 13:46:29 crc kubenswrapper[4719]: I1215 13:46:29.895271 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-55f78b7c4c-292vp_a9ebdf36-2095-47b8-8edf-89036ffdee84/manager/0.log" Dec 15 13:46:31 crc kubenswrapper[4719]: I1215 13:46:31.476377 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:46:31 crc kubenswrapper[4719]: E1215 13:46:31.476920 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:46:46 crc kubenswrapper[4719]: I1215 13:46:46.476683 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:46:46 crc kubenswrapper[4719]: E1215 13:46:46.477518 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:46:53 crc kubenswrapper[4719]: I1215 13:46:53.035464 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ngv2q_d620f9fd-1374-4f93-972d-93f2b9173471/control-plane-machine-set-operator/0.log" Dec 15 13:46:53 crc kubenswrapper[4719]: I1215 13:46:53.381025 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c5df6_0dd6e306-1355-43dd-895f-598d3d873a42/kube-rbac-proxy/0.log" Dec 15 13:46:53 crc kubenswrapper[4719]: I1215 13:46:53.456661 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c5df6_0dd6e306-1355-43dd-895f-598d3d873a42/machine-api-operator/0.log" Dec 15 13:47:01 crc kubenswrapper[4719]: I1215 13:47:01.479478 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:47:02 crc kubenswrapper[4719]: I1215 13:47:02.670135 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"d07a1864400809645dee53e0d4f9f1e7b25073a20cbc7b6da401e405d87bc912"} Dec 15 13:47:06 crc kubenswrapper[4719]: I1215 13:47:06.540198 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-6mllc_87e97929-9f4b-4e2e-bbc2-e9295954a8ae/cert-manager-controller/0.log" Dec 15 13:47:06 crc kubenswrapper[4719]: I1215 13:47:06.747796 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-pjzlc_72cfd6e6-a544-4002-81ab-b57dfd2eaf94/cert-manager-cainjector/0.log" Dec 15 13:47:06 crc kubenswrapper[4719]: I1215 13:47:06.827347 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-nkj5k_994c9412-da12-497c-8849-a5948a36e975/cert-manager-webhook/0.log" Dec 15 13:47:18 crc kubenswrapper[4719]: I1215 13:47:18.863405 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6ff7998486-ljnvk_5b49fba8-f261-4a35-9aa0-669b7e7e70ab/nmstate-console-plugin/0.log" Dec 15 13:47:19 crc kubenswrapper[4719]: I1215 13:47:19.051412 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-trxk8_6524d6cd-421e-415d-826b-7ec166bc3915/nmstate-handler/0.log" Dec 15 13:47:19 crc kubenswrapper[4719]: I1215 13:47:19.165330 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-vwmdc_1fd5707c-b363-45cc-b2b3-168817ec1526/kube-rbac-proxy/0.log" Dec 15 13:47:19 crc kubenswrapper[4719]: I1215 13:47:19.249151 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-vwmdc_1fd5707c-b363-45cc-b2b3-168817ec1526/nmstate-metrics/0.log" Dec 15 13:47:19 crc kubenswrapper[4719]: I1215 13:47:19.381635 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-6769fb99d-h4ptq_91b08102-b558-4d78-ade6-0360f6e7846a/nmstate-operator/0.log" Dec 15 13:47:19 crc kubenswrapper[4719]: I1215 13:47:19.492373 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-f8fb84555-kqgm4_cee112f9-d46f-410b-bfc5-5d6cff881bb4/nmstate-webhook/0.log" Dec 15 13:47:34 crc kubenswrapper[4719]: I1215 13:47:34.698835 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-96g4h_49178e96-75a4-4b6d-8540-781182eb8123/controller/0.log" Dec 15 13:47:34 crc kubenswrapper[4719]: I1215 13:47:34.729997 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-96g4h_49178e96-75a4-4b6d-8540-781182eb8123/kube-rbac-proxy/0.log" Dec 15 13:47:34 crc kubenswrapper[4719]: I1215 13:47:34.954362 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-frr-files/0.log" Dec 15 13:47:34 crc kubenswrapper[4719]: I1215 13:47:34.969911 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-crztf_6bac73e3-46cd-4cca-bb26-1ae16f1c804e/frr-k8s-webhook-server/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.144754 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-frr-files/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.162302 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-reloader/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.222701 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-metrics/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.236721 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-reloader/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.397439 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-frr-files/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.398438 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-reloader/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.477972 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-metrics/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.492610 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-metrics/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.611971 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-frr-files/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.674440 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-reloader/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.737236 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/controller/0.log" Dec 15 13:47:35 crc kubenswrapper[4719]: I1215 13:47:35.756199 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/cp-metrics/0.log" Dec 15 13:47:36 crc kubenswrapper[4719]: I1215 13:47:36.005203 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/frr-metrics/0.log" Dec 15 13:47:36 crc kubenswrapper[4719]: I1215 13:47:36.064255 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/kube-rbac-proxy/0.log" Dec 15 13:47:36 crc kubenswrapper[4719]: I1215 13:47:36.090222 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/kube-rbac-proxy-frr/0.log" Dec 15 13:47:36 crc kubenswrapper[4719]: I1215 13:47:36.429340 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/reloader/0.log" Dec 15 13:47:36 crc kubenswrapper[4719]: I1215 13:47:36.476645 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6cf5cc9b68-dpxck_5009ea4d-5e13-4ac3-ad25-8c0976b0aafb/manager/0.log" Dec 15 13:47:36 crc kubenswrapper[4719]: I1215 13:47:36.735758 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-76fbf99b47-d626r_15904dfe-b1ca-4922-95cf-6f3724fae181/webhook-server/0.log" Dec 15 13:47:36 crc kubenswrapper[4719]: I1215 13:47:36.921379 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tdxf5_72e46a3d-9b79-404d-80ba-68a363c23b2b/kube-rbac-proxy/0.log" Dec 15 13:47:37 crc kubenswrapper[4719]: I1215 13:47:37.476640 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tdxf5_72e46a3d-9b79-404d-80ba-68a363c23b2b/speaker/0.log" Dec 15 13:47:37 crc kubenswrapper[4719]: I1215 13:47:37.515425 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-wnthh_dbea97f4-325a-4d31-8562-c1f5b546fca5/frr/0.log" Dec 15 13:47:50 crc kubenswrapper[4719]: I1215 13:47:50.992795 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/util/0.log" Dec 15 13:47:51 crc kubenswrapper[4719]: I1215 13:47:51.183089 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/util/0.log" Dec 15 13:47:51 crc kubenswrapper[4719]: I1215 13:47:51.260397 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/pull/0.log" Dec 15 13:47:51 crc kubenswrapper[4719]: I1215 13:47:51.274918 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/pull/0.log" Dec 15 13:47:51 crc kubenswrapper[4719]: I1215 13:47:51.510581 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/util/0.log" Dec 15 13:47:51 crc kubenswrapper[4719]: I1215 13:47:51.544328 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/pull/0.log" Dec 15 13:47:51 crc kubenswrapper[4719]: I1215 13:47:51.563964 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d46jhmr_2de0d9f9-5f17-47b6-ab6e-c48ce52f8db0/extract/0.log" Dec 15 13:47:51 crc kubenswrapper[4719]: I1215 13:47:51.738816 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/util/0.log" Dec 15 13:47:51 crc kubenswrapper[4719]: I1215 13:47:51.941520 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/util/0.log" Dec 15 13:47:51 crc kubenswrapper[4719]: I1215 13:47:51.983160 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/pull/0.log" Dec 15 13:47:52 crc kubenswrapper[4719]: I1215 13:47:52.032367 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/pull/0.log" Dec 15 13:47:52 crc kubenswrapper[4719]: I1215 13:47:52.218219 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/util/0.log" Dec 15 13:47:52 crc kubenswrapper[4719]: I1215 13:47:52.218574 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/pull/0.log" Dec 15 13:47:52 crc kubenswrapper[4719]: I1215 13:47:52.241714 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8tq56m_6def757d-1beb-4122-af03-45d75afd2574/extract/0.log" Dec 15 13:47:52 crc kubenswrapper[4719]: I1215 13:47:52.439945 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-utilities/0.log" Dec 15 13:47:52 crc kubenswrapper[4719]: I1215 13:47:52.693102 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-content/0.log" Dec 15 13:47:52 crc kubenswrapper[4719]: I1215 13:47:52.720296 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-utilities/0.log" Dec 15 13:47:52 crc kubenswrapper[4719]: I1215 13:47:52.731485 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-content/0.log" Dec 15 13:47:52 crc kubenswrapper[4719]: I1215 13:47:52.886140 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-content/0.log" Dec 15 13:47:52 crc kubenswrapper[4719]: I1215 13:47:52.904549 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/extract-utilities/0.log" Dec 15 13:47:53 crc kubenswrapper[4719]: I1215 13:47:53.224216 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-utilities/0.log" Dec 15 13:47:53 crc kubenswrapper[4719]: I1215 13:47:53.409902 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-utilities/0.log" Dec 15 13:47:53 crc kubenswrapper[4719]: I1215 13:47:53.573949 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-content/0.log" Dec 15 13:47:53 crc kubenswrapper[4719]: I1215 13:47:53.587425 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lhrf4_ea4a4126-7481-417c-968b-f3f4e0b5ff6e/registry-server/0.log" Dec 15 13:47:53 crc kubenswrapper[4719]: I1215 13:47:53.638499 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-content/0.log" Dec 15 13:47:53 crc kubenswrapper[4719]: I1215 13:47:53.854571 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-content/0.log" Dec 15 13:47:53 crc kubenswrapper[4719]: I1215 13:47:53.935570 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/extract-utilities/0.log" Dec 15 13:47:54 crc kubenswrapper[4719]: I1215 13:47:54.186744 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kxrn8_e804e02c-d21b-4881-9bc8-2de04a0928f9/marketplace-operator/0.log" Dec 15 13:47:54 crc kubenswrapper[4719]: I1215 13:47:54.486810 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sgc7g_9d9caada-f1dc-4aa5-b328-1e1442300524/registry-server/0.log" Dec 15 13:47:54 crc kubenswrapper[4719]: I1215 13:47:54.517659 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-utilities/0.log" Dec 15 13:47:54 crc kubenswrapper[4719]: I1215 13:47:54.948276 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-utilities/0.log" Dec 15 13:47:55 crc kubenswrapper[4719]: I1215 13:47:55.017072 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-content/0.log" Dec 15 13:47:55 crc kubenswrapper[4719]: I1215 13:47:55.037728 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-content/0.log" Dec 15 13:47:55 crc kubenswrapper[4719]: I1215 13:47:55.188630 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-content/0.log" Dec 15 13:47:55 crc kubenswrapper[4719]: I1215 13:47:55.194644 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/extract-utilities/0.log" Dec 15 13:47:55 crc kubenswrapper[4719]: I1215 13:47:55.408177 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nrldb_577da6cd-6fcd-424a-8d6c-2aa4fe6e4c2b/registry-server/0.log" Dec 15 13:47:55 crc kubenswrapper[4719]: I1215 13:47:55.475165 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-utilities/0.log" Dec 15 13:47:55 crc kubenswrapper[4719]: I1215 13:47:55.699542 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-utilities/0.log" Dec 15 13:47:55 crc kubenswrapper[4719]: I1215 13:47:55.707496 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-content/0.log" Dec 15 13:47:55 crc kubenswrapper[4719]: I1215 13:47:55.777572 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-content/0.log" Dec 15 13:47:55 crc kubenswrapper[4719]: I1215 13:47:55.991516 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-utilities/0.log" Dec 15 13:47:56 crc kubenswrapper[4719]: I1215 13:47:56.042316 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/extract-content/0.log" Dec 15 13:47:56 crc kubenswrapper[4719]: I1215 13:47:56.238584 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j7xqn_6aed14ad-077e-449b-a183-d583dc39893d/registry-server/0.log" Dec 15 13:48:01 crc kubenswrapper[4719]: I1215 13:48:01.823106 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7jrg6"] Dec 15 13:48:01 crc kubenswrapper[4719]: E1215 13:48:01.824117 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfd55b12-b976-45f5-b07f-abcb95a350e3" containerName="container-00" Dec 15 13:48:01 crc kubenswrapper[4719]: I1215 13:48:01.824133 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfd55b12-b976-45f5-b07f-abcb95a350e3" containerName="container-00" Dec 15 13:48:01 crc kubenswrapper[4719]: I1215 13:48:01.824387 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfd55b12-b976-45f5-b07f-abcb95a350e3" containerName="container-00" Dec 15 13:48:01 crc kubenswrapper[4719]: I1215 13:48:01.826548 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:01 crc kubenswrapper[4719]: I1215 13:48:01.838477 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7jrg6"] Dec 15 13:48:01 crc kubenswrapper[4719]: I1215 13:48:01.996362 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-catalog-content\") pod \"certified-operators-7jrg6\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:01 crc kubenswrapper[4719]: I1215 13:48:01.996682 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2zcc\" (UniqueName: \"kubernetes.io/projected/18535fe9-0a1b-463e-a916-61e2855c52fe-kube-api-access-f2zcc\") pod \"certified-operators-7jrg6\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:01 crc kubenswrapper[4719]: I1215 13:48:01.996806 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-utilities\") pod \"certified-operators-7jrg6\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:02 crc kubenswrapper[4719]: I1215 13:48:02.098544 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-catalog-content\") pod \"certified-operators-7jrg6\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:02 crc kubenswrapper[4719]: I1215 13:48:02.098597 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2zcc\" (UniqueName: \"kubernetes.io/projected/18535fe9-0a1b-463e-a916-61e2855c52fe-kube-api-access-f2zcc\") pod \"certified-operators-7jrg6\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:02 crc kubenswrapper[4719]: I1215 13:48:02.098640 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-utilities\") pod \"certified-operators-7jrg6\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:02 crc kubenswrapper[4719]: I1215 13:48:02.099100 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-catalog-content\") pod \"certified-operators-7jrg6\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:02 crc kubenswrapper[4719]: I1215 13:48:02.099142 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-utilities\") pod \"certified-operators-7jrg6\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:02 crc kubenswrapper[4719]: I1215 13:48:02.128982 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2zcc\" (UniqueName: \"kubernetes.io/projected/18535fe9-0a1b-463e-a916-61e2855c52fe-kube-api-access-f2zcc\") pod \"certified-operators-7jrg6\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:02 crc kubenswrapper[4719]: I1215 13:48:02.153877 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:02 crc kubenswrapper[4719]: I1215 13:48:02.721556 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7jrg6"] Dec 15 13:48:03 crc kubenswrapper[4719]: I1215 13:48:03.185276 4719 generic.go:334] "Generic (PLEG): container finished" podID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerID="1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6" exitCode=0 Dec 15 13:48:03 crc kubenswrapper[4719]: I1215 13:48:03.185390 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jrg6" event={"ID":"18535fe9-0a1b-463e-a916-61e2855c52fe","Type":"ContainerDied","Data":"1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6"} Dec 15 13:48:03 crc kubenswrapper[4719]: I1215 13:48:03.185593 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jrg6" event={"ID":"18535fe9-0a1b-463e-a916-61e2855c52fe","Type":"ContainerStarted","Data":"0f6ad742c34bbe58e927bd3788a4f7f4dc516ce4fb6eab2c5ec2797f6a2fbef7"} Dec 15 13:48:05 crc kubenswrapper[4719]: I1215 13:48:05.217649 4719 generic.go:334] "Generic (PLEG): container finished" podID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerID="149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3" exitCode=0 Dec 15 13:48:05 crc kubenswrapper[4719]: I1215 13:48:05.218216 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jrg6" event={"ID":"18535fe9-0a1b-463e-a916-61e2855c52fe","Type":"ContainerDied","Data":"149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3"} Dec 15 13:48:06 crc kubenswrapper[4719]: I1215 13:48:06.227414 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jrg6" event={"ID":"18535fe9-0a1b-463e-a916-61e2855c52fe","Type":"ContainerStarted","Data":"89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded"} Dec 15 13:48:06 crc kubenswrapper[4719]: I1215 13:48:06.250062 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7jrg6" podStartSLOduration=2.762193537 podStartE2EDuration="5.250039629s" podCreationTimestamp="2025-12-15 13:48:01 +0000 UTC" firstStartedPulling="2025-12-15 13:48:03.187786381 +0000 UTC m=+5444.130079411" lastFinishedPulling="2025-12-15 13:48:05.675632463 +0000 UTC m=+5446.617925503" observedRunningTime="2025-12-15 13:48:06.241468479 +0000 UTC m=+5447.183761519" watchObservedRunningTime="2025-12-15 13:48:06.250039629 +0000 UTC m=+5447.192332659" Dec 15 13:48:12 crc kubenswrapper[4719]: I1215 13:48:12.155357 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:12 crc kubenswrapper[4719]: I1215 13:48:12.156185 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:12 crc kubenswrapper[4719]: I1215 13:48:12.206848 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:12 crc kubenswrapper[4719]: I1215 13:48:12.339803 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:12 crc kubenswrapper[4719]: I1215 13:48:12.441491 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7jrg6"] Dec 15 13:48:14 crc kubenswrapper[4719]: I1215 13:48:14.289044 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7jrg6" podUID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerName="registry-server" containerID="cri-o://89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded" gracePeriod=2 Dec 15 13:48:14 crc kubenswrapper[4719]: I1215 13:48:14.854282 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:14 crc kubenswrapper[4719]: I1215 13:48:14.959188 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-utilities\") pod \"18535fe9-0a1b-463e-a916-61e2855c52fe\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " Dec 15 13:48:14 crc kubenswrapper[4719]: I1215 13:48:14.959389 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2zcc\" (UniqueName: \"kubernetes.io/projected/18535fe9-0a1b-463e-a916-61e2855c52fe-kube-api-access-f2zcc\") pod \"18535fe9-0a1b-463e-a916-61e2855c52fe\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " Dec 15 13:48:14 crc kubenswrapper[4719]: I1215 13:48:14.959482 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-catalog-content\") pod \"18535fe9-0a1b-463e-a916-61e2855c52fe\" (UID: \"18535fe9-0a1b-463e-a916-61e2855c52fe\") " Dec 15 13:48:14 crc kubenswrapper[4719]: I1215 13:48:14.960388 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-utilities" (OuterVolumeSpecName: "utilities") pod "18535fe9-0a1b-463e-a916-61e2855c52fe" (UID: "18535fe9-0a1b-463e-a916-61e2855c52fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:48:14 crc kubenswrapper[4719]: I1215 13:48:14.983837 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18535fe9-0a1b-463e-a916-61e2855c52fe-kube-api-access-f2zcc" (OuterVolumeSpecName: "kube-api-access-f2zcc") pod "18535fe9-0a1b-463e-a916-61e2855c52fe" (UID: "18535fe9-0a1b-463e-a916-61e2855c52fe"). InnerVolumeSpecName "kube-api-access-f2zcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.042691 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18535fe9-0a1b-463e-a916-61e2855c52fe" (UID: "18535fe9-0a1b-463e-a916-61e2855c52fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.061569 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.061604 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18535fe9-0a1b-463e-a916-61e2855c52fe-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.061614 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2zcc\" (UniqueName: \"kubernetes.io/projected/18535fe9-0a1b-463e-a916-61e2855c52fe-kube-api-access-f2zcc\") on node \"crc\" DevicePath \"\"" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.299934 4719 generic.go:334] "Generic (PLEG): container finished" podID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerID="89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded" exitCode=0 Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.299984 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jrg6" event={"ID":"18535fe9-0a1b-463e-a916-61e2855c52fe","Type":"ContainerDied","Data":"89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded"} Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.301070 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jrg6" event={"ID":"18535fe9-0a1b-463e-a916-61e2855c52fe","Type":"ContainerDied","Data":"0f6ad742c34bbe58e927bd3788a4f7f4dc516ce4fb6eab2c5ec2797f6a2fbef7"} Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.300004 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jrg6" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.301132 4719 scope.go:117] "RemoveContainer" containerID="89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.326912 4719 scope.go:117] "RemoveContainer" containerID="149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.345151 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7jrg6"] Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.354047 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7jrg6"] Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.371950 4719 scope.go:117] "RemoveContainer" containerID="1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.411805 4719 scope.go:117] "RemoveContainer" containerID="89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded" Dec 15 13:48:15 crc kubenswrapper[4719]: E1215 13:48:15.412312 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded\": container with ID starting with 89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded not found: ID does not exist" containerID="89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.412336 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded"} err="failed to get container status \"89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded\": rpc error: code = NotFound desc = could not find container \"89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded\": container with ID starting with 89c12f875c61cf8c1620d5b8cd1094f628598b98e5dfb94d0534837f40426ded not found: ID does not exist" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.412356 4719 scope.go:117] "RemoveContainer" containerID="149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3" Dec 15 13:48:15 crc kubenswrapper[4719]: E1215 13:48:15.412644 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3\": container with ID starting with 149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3 not found: ID does not exist" containerID="149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.412658 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3"} err="failed to get container status \"149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3\": rpc error: code = NotFound desc = could not find container \"149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3\": container with ID starting with 149d048d5fc0a798e3ad1e7712323a94572d184f6a9a15a03330906afd422ef3 not found: ID does not exist" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.412670 4719 scope.go:117] "RemoveContainer" containerID="1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6" Dec 15 13:48:15 crc kubenswrapper[4719]: E1215 13:48:15.412842 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6\": container with ID starting with 1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6 not found: ID does not exist" containerID="1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.412871 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6"} err="failed to get container status \"1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6\": rpc error: code = NotFound desc = could not find container \"1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6\": container with ID starting with 1fa115129cc67a59c0104ddcbd702c8c445716a29118d0d2286aa87890af4ee6 not found: ID does not exist" Dec 15 13:48:15 crc kubenswrapper[4719]: I1215 13:48:15.496246 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18535fe9-0a1b-463e-a916-61e2855c52fe" path="/var/lib/kubelet/pods/18535fe9-0a1b-463e-a916-61e2855c52fe/volumes" Dec 15 13:48:15 crc kubenswrapper[4719]: E1215 13:48:15.550047 4719 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.217:41054->38.102.83.217:45485: read tcp 38.102.83.217:41054->38.102.83.217:45485: read: connection reset by peer Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.457649 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4pdk8"] Dec 15 13:48:27 crc kubenswrapper[4719]: E1215 13:48:27.458650 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerName="extract-content" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.458665 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerName="extract-content" Dec 15 13:48:27 crc kubenswrapper[4719]: E1215 13:48:27.458699 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerName="registry-server" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.458708 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerName="registry-server" Dec 15 13:48:27 crc kubenswrapper[4719]: E1215 13:48:27.458729 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerName="extract-utilities" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.458738 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerName="extract-utilities" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.458995 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="18535fe9-0a1b-463e-a916-61e2855c52fe" containerName="registry-server" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.460592 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.473842 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pdk8"] Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.618262 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-utilities\") pod \"redhat-marketplace-4pdk8\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.618369 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-catalog-content\") pod \"redhat-marketplace-4pdk8\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.618528 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbpzv\" (UniqueName: \"kubernetes.io/projected/d103e4af-ade5-44e3-9608-06e3e7fff3fc-kube-api-access-cbpzv\") pod \"redhat-marketplace-4pdk8\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.722051 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbpzv\" (UniqueName: \"kubernetes.io/projected/d103e4af-ade5-44e3-9608-06e3e7fff3fc-kube-api-access-cbpzv\") pod \"redhat-marketplace-4pdk8\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.722187 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-utilities\") pod \"redhat-marketplace-4pdk8\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.722254 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-catalog-content\") pod \"redhat-marketplace-4pdk8\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.722894 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-catalog-content\") pod \"redhat-marketplace-4pdk8\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.723173 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-utilities\") pod \"redhat-marketplace-4pdk8\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.745227 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbpzv\" (UniqueName: \"kubernetes.io/projected/d103e4af-ade5-44e3-9608-06e3e7fff3fc-kube-api-access-cbpzv\") pod \"redhat-marketplace-4pdk8\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:27 crc kubenswrapper[4719]: I1215 13:48:27.780600 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:28 crc kubenswrapper[4719]: I1215 13:48:28.369918 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pdk8"] Dec 15 13:48:28 crc kubenswrapper[4719]: I1215 13:48:28.414281 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pdk8" event={"ID":"d103e4af-ade5-44e3-9608-06e3e7fff3fc","Type":"ContainerStarted","Data":"0bf95f1bdd6d64d20b40f549eb73256ab97cc372a010ad0d5261473b16fba5fe"} Dec 15 13:48:29 crc kubenswrapper[4719]: I1215 13:48:29.423082 4719 generic.go:334] "Generic (PLEG): container finished" podID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerID="163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5" exitCode=0 Dec 15 13:48:29 crc kubenswrapper[4719]: I1215 13:48:29.423297 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pdk8" event={"ID":"d103e4af-ade5-44e3-9608-06e3e7fff3fc","Type":"ContainerDied","Data":"163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5"} Dec 15 13:48:30 crc kubenswrapper[4719]: I1215 13:48:30.432741 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pdk8" event={"ID":"d103e4af-ade5-44e3-9608-06e3e7fff3fc","Type":"ContainerStarted","Data":"3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb"} Dec 15 13:48:31 crc kubenswrapper[4719]: I1215 13:48:31.446226 4719 generic.go:334] "Generic (PLEG): container finished" podID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerID="3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb" exitCode=0 Dec 15 13:48:31 crc kubenswrapper[4719]: I1215 13:48:31.446265 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pdk8" event={"ID":"d103e4af-ade5-44e3-9608-06e3e7fff3fc","Type":"ContainerDied","Data":"3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb"} Dec 15 13:48:32 crc kubenswrapper[4719]: I1215 13:48:32.485317 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pdk8" event={"ID":"d103e4af-ade5-44e3-9608-06e3e7fff3fc","Type":"ContainerStarted","Data":"c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314"} Dec 15 13:48:32 crc kubenswrapper[4719]: I1215 13:48:32.526237 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4pdk8" podStartSLOduration=2.858629695 podStartE2EDuration="5.526204927s" podCreationTimestamp="2025-12-15 13:48:27 +0000 UTC" firstStartedPulling="2025-12-15 13:48:29.425040296 +0000 UTC m=+5470.367333326" lastFinishedPulling="2025-12-15 13:48:32.092615538 +0000 UTC m=+5473.034908558" observedRunningTime="2025-12-15 13:48:32.515175551 +0000 UTC m=+5473.457468581" watchObservedRunningTime="2025-12-15 13:48:32.526204927 +0000 UTC m=+5473.468497977" Dec 15 13:48:37 crc kubenswrapper[4719]: I1215 13:48:37.781868 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:37 crc kubenswrapper[4719]: I1215 13:48:37.782461 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:37 crc kubenswrapper[4719]: I1215 13:48:37.839715 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:38 crc kubenswrapper[4719]: I1215 13:48:38.621810 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:38 crc kubenswrapper[4719]: I1215 13:48:38.701982 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pdk8"] Dec 15 13:48:40 crc kubenswrapper[4719]: I1215 13:48:40.544155 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4pdk8" podUID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerName="registry-server" containerID="cri-o://c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314" gracePeriod=2 Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.062121 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.228824 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-catalog-content\") pod \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.229052 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-utilities\") pod \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.229137 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbpzv\" (UniqueName: \"kubernetes.io/projected/d103e4af-ade5-44e3-9608-06e3e7fff3fc-kube-api-access-cbpzv\") pod \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\" (UID: \"d103e4af-ade5-44e3-9608-06e3e7fff3fc\") " Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.230519 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-utilities" (OuterVolumeSpecName: "utilities") pod "d103e4af-ade5-44e3-9608-06e3e7fff3fc" (UID: "d103e4af-ade5-44e3-9608-06e3e7fff3fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.235002 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d103e4af-ade5-44e3-9608-06e3e7fff3fc-kube-api-access-cbpzv" (OuterVolumeSpecName: "kube-api-access-cbpzv") pod "d103e4af-ade5-44e3-9608-06e3e7fff3fc" (UID: "d103e4af-ade5-44e3-9608-06e3e7fff3fc"). InnerVolumeSpecName "kube-api-access-cbpzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.255452 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d103e4af-ade5-44e3-9608-06e3e7fff3fc" (UID: "d103e4af-ade5-44e3-9608-06e3e7fff3fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.331835 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.331891 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d103e4af-ade5-44e3-9608-06e3e7fff3fc-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.331902 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbpzv\" (UniqueName: \"kubernetes.io/projected/d103e4af-ade5-44e3-9608-06e3e7fff3fc-kube-api-access-cbpzv\") on node \"crc\" DevicePath \"\"" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.556329 4719 generic.go:334] "Generic (PLEG): container finished" podID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerID="c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314" exitCode=0 Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.556376 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pdk8" event={"ID":"d103e4af-ade5-44e3-9608-06e3e7fff3fc","Type":"ContainerDied","Data":"c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314"} Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.556408 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pdk8" event={"ID":"d103e4af-ade5-44e3-9608-06e3e7fff3fc","Type":"ContainerDied","Data":"0bf95f1bdd6d64d20b40f549eb73256ab97cc372a010ad0d5261473b16fba5fe"} Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.556432 4719 scope.go:117] "RemoveContainer" containerID="c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.556584 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pdk8" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.592974 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pdk8"] Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.597798 4719 scope.go:117] "RemoveContainer" containerID="3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.634891 4719 scope.go:117] "RemoveContainer" containerID="163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.655664 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pdk8"] Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.703475 4719 scope.go:117] "RemoveContainer" containerID="c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314" Dec 15 13:48:41 crc kubenswrapper[4719]: E1215 13:48:41.704345 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314\": container with ID starting with c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314 not found: ID does not exist" containerID="c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.704403 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314"} err="failed to get container status \"c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314\": rpc error: code = NotFound desc = could not find container \"c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314\": container with ID starting with c3184c4e954489478582915af0f9a6c467e46aa58c1a9816a5505523f7fd3314 not found: ID does not exist" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.704442 4719 scope.go:117] "RemoveContainer" containerID="3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb" Dec 15 13:48:41 crc kubenswrapper[4719]: E1215 13:48:41.705016 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb\": container with ID starting with 3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb not found: ID does not exist" containerID="3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.705073 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb"} err="failed to get container status \"3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb\": rpc error: code = NotFound desc = could not find container \"3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb\": container with ID starting with 3b703ba679487cb462be63c6f404b15f74fa391891d9d7fff95724225154eadb not found: ID does not exist" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.705110 4719 scope.go:117] "RemoveContainer" containerID="163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5" Dec 15 13:48:41 crc kubenswrapper[4719]: E1215 13:48:41.705755 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5\": container with ID starting with 163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5 not found: ID does not exist" containerID="163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5" Dec 15 13:48:41 crc kubenswrapper[4719]: I1215 13:48:41.705792 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5"} err="failed to get container status \"163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5\": rpc error: code = NotFound desc = could not find container \"163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5\": container with ID starting with 163d470be52c29fc754589650ff01e7659fafbc5f73bfec8d0a510874c2abfd5 not found: ID does not exist" Dec 15 13:48:43 crc kubenswrapper[4719]: I1215 13:48:43.491341 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" path="/var/lib/kubelet/pods/d103e4af-ade5-44e3-9608-06e3e7fff3fc/volumes" Dec 15 13:49:21 crc kubenswrapper[4719]: I1215 13:49:21.357015 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:49:21 crc kubenswrapper[4719]: I1215 13:49:21.357596 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:49:51 crc kubenswrapper[4719]: I1215 13:49:51.356917 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:49:51 crc kubenswrapper[4719]: I1215 13:49:51.358691 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:50:16 crc kubenswrapper[4719]: I1215 13:50:16.380661 4719 generic.go:334] "Generic (PLEG): container finished" podID="f3679f93-2fdb-4155-be86-eac3670559ea" containerID="f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648" exitCode=0 Dec 15 13:50:16 crc kubenswrapper[4719]: I1215 13:50:16.380739 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bf8hx/must-gather-xzsl8" event={"ID":"f3679f93-2fdb-4155-be86-eac3670559ea","Type":"ContainerDied","Data":"f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648"} Dec 15 13:50:16 crc kubenswrapper[4719]: I1215 13:50:16.381991 4719 scope.go:117] "RemoveContainer" containerID="f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648" Dec 15 13:50:16 crc kubenswrapper[4719]: I1215 13:50:16.838405 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bf8hx_must-gather-xzsl8_f3679f93-2fdb-4155-be86-eac3670559ea/gather/0.log" Dec 15 13:50:21 crc kubenswrapper[4719]: I1215 13:50:21.356643 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:50:21 crc kubenswrapper[4719]: I1215 13:50:21.358725 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:50:21 crc kubenswrapper[4719]: I1215 13:50:21.358769 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 13:50:21 crc kubenswrapper[4719]: I1215 13:50:21.359302 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d07a1864400809645dee53e0d4f9f1e7b25073a20cbc7b6da401e405d87bc912"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 13:50:21 crc kubenswrapper[4719]: I1215 13:50:21.359349 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://d07a1864400809645dee53e0d4f9f1e7b25073a20cbc7b6da401e405d87bc912" gracePeriod=600 Dec 15 13:50:22 crc kubenswrapper[4719]: I1215 13:50:22.438935 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="d07a1864400809645dee53e0d4f9f1e7b25073a20cbc7b6da401e405d87bc912" exitCode=0 Dec 15 13:50:22 crc kubenswrapper[4719]: I1215 13:50:22.438986 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"d07a1864400809645dee53e0d4f9f1e7b25073a20cbc7b6da401e405d87bc912"} Dec 15 13:50:22 crc kubenswrapper[4719]: I1215 13:50:22.439581 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerStarted","Data":"7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7"} Dec 15 13:50:22 crc kubenswrapper[4719]: I1215 13:50:22.439605 4719 scope.go:117] "RemoveContainer" containerID="6b66c8e9c8897d1413f9ca099622727eeffce629192a920801dc7026db20d04a" Dec 15 13:50:29 crc kubenswrapper[4719]: I1215 13:50:29.884390 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bf8hx/must-gather-xzsl8"] Dec 15 13:50:29 crc kubenswrapper[4719]: I1215 13:50:29.885242 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-bf8hx/must-gather-xzsl8" podUID="f3679f93-2fdb-4155-be86-eac3670559ea" containerName="copy" containerID="cri-o://df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c" gracePeriod=2 Dec 15 13:50:29 crc kubenswrapper[4719]: I1215 13:50:29.896494 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bf8hx/must-gather-xzsl8"] Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.294384 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bf8hx_must-gather-xzsl8_f3679f93-2fdb-4155-be86-eac3670559ea/copy/0.log" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.295034 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/must-gather-xzsl8" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.439613 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6mp8\" (UniqueName: \"kubernetes.io/projected/f3679f93-2fdb-4155-be86-eac3670559ea-kube-api-access-f6mp8\") pod \"f3679f93-2fdb-4155-be86-eac3670559ea\" (UID: \"f3679f93-2fdb-4155-be86-eac3670559ea\") " Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.439770 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f3679f93-2fdb-4155-be86-eac3670559ea-must-gather-output\") pod \"f3679f93-2fdb-4155-be86-eac3670559ea\" (UID: \"f3679f93-2fdb-4155-be86-eac3670559ea\") " Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.450626 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3679f93-2fdb-4155-be86-eac3670559ea-kube-api-access-f6mp8" (OuterVolumeSpecName: "kube-api-access-f6mp8") pod "f3679f93-2fdb-4155-be86-eac3670559ea" (UID: "f3679f93-2fdb-4155-be86-eac3670559ea"). InnerVolumeSpecName "kube-api-access-f6mp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.519901 4719 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bf8hx_must-gather-xzsl8_f3679f93-2fdb-4155-be86-eac3670559ea/copy/0.log" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.520204 4719 generic.go:334] "Generic (PLEG): container finished" podID="f3679f93-2fdb-4155-be86-eac3670559ea" containerID="df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c" exitCode=143 Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.520257 4719 scope.go:117] "RemoveContainer" containerID="df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.520407 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bf8hx/must-gather-xzsl8" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.543074 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6mp8\" (UniqueName: \"kubernetes.io/projected/f3679f93-2fdb-4155-be86-eac3670559ea-kube-api-access-f6mp8\") on node \"crc\" DevicePath \"\"" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.579619 4719 scope.go:117] "RemoveContainer" containerID="f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.676234 4719 scope.go:117] "RemoveContainer" containerID="df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c" Dec 15 13:50:30 crc kubenswrapper[4719]: E1215 13:50:30.678581 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c\": container with ID starting with df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c not found: ID does not exist" containerID="df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.678640 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c"} err="failed to get container status \"df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c\": rpc error: code = NotFound desc = could not find container \"df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c\": container with ID starting with df81acb6482a922ee04628a51015d57ef245e5c9939dbeb212401a5baf07078c not found: ID does not exist" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.678668 4719 scope.go:117] "RemoveContainer" containerID="f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648" Dec 15 13:50:30 crc kubenswrapper[4719]: E1215 13:50:30.681683 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648\": container with ID starting with f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648 not found: ID does not exist" containerID="f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.681740 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648"} err="failed to get container status \"f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648\": rpc error: code = NotFound desc = could not find container \"f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648\": container with ID starting with f67fcb565a6a36762fcb2094b4e3689e0e62768a0bad5b0bed5c38a70c673648 not found: ID does not exist" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.713545 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3679f93-2fdb-4155-be86-eac3670559ea-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "f3679f93-2fdb-4155-be86-eac3670559ea" (UID: "f3679f93-2fdb-4155-be86-eac3670559ea"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:50:30 crc kubenswrapper[4719]: I1215 13:50:30.750288 4719 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f3679f93-2fdb-4155-be86-eac3670559ea-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 15 13:50:31 crc kubenswrapper[4719]: I1215 13:50:31.491080 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3679f93-2fdb-4155-be86-eac3670559ea" path="/var/lib/kubelet/pods/f3679f93-2fdb-4155-be86-eac3670559ea/volumes" Dec 15 13:51:36 crc kubenswrapper[4719]: I1215 13:51:36.673758 4719 scope.go:117] "RemoveContainer" containerID="b7025df69cbb1837897f9c27b60cdcd8ea3628e1e45d299c0d124e7ce41f9a5f" Dec 15 13:52:21 crc kubenswrapper[4719]: I1215 13:52:21.357474 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:52:21 crc kubenswrapper[4719]: I1215 13:52:21.358415 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.797976 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nwgpc"] Dec 15 13:52:26 crc kubenswrapper[4719]: E1215 13:52:26.800032 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerName="extract-content" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.800145 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerName="extract-content" Dec 15 13:52:26 crc kubenswrapper[4719]: E1215 13:52:26.800245 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3679f93-2fdb-4155-be86-eac3670559ea" containerName="gather" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.800302 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3679f93-2fdb-4155-be86-eac3670559ea" containerName="gather" Dec 15 13:52:26 crc kubenswrapper[4719]: E1215 13:52:26.800369 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerName="registry-server" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.800422 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerName="registry-server" Dec 15 13:52:26 crc kubenswrapper[4719]: E1215 13:52:26.800492 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3679f93-2fdb-4155-be86-eac3670559ea" containerName="copy" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.800567 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3679f93-2fdb-4155-be86-eac3670559ea" containerName="copy" Dec 15 13:52:26 crc kubenswrapper[4719]: E1215 13:52:26.800630 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerName="extract-utilities" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.800685 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerName="extract-utilities" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.800965 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="d103e4af-ade5-44e3-9608-06e3e7fff3fc" containerName="registry-server" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.801040 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3679f93-2fdb-4155-be86-eac3670559ea" containerName="gather" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.801178 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3679f93-2fdb-4155-be86-eac3670559ea" containerName="copy" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.802660 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.815894 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwgpc"] Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.836345 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-catalog-content\") pod \"community-operators-nwgpc\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.836451 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qccld\" (UniqueName: \"kubernetes.io/projected/c5111a80-08c9-4933-8a65-906b58149ade-kube-api-access-qccld\") pod \"community-operators-nwgpc\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.836489 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-utilities\") pod \"community-operators-nwgpc\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.938288 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qccld\" (UniqueName: \"kubernetes.io/projected/c5111a80-08c9-4933-8a65-906b58149ade-kube-api-access-qccld\") pod \"community-operators-nwgpc\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.938344 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-utilities\") pod \"community-operators-nwgpc\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.938424 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-catalog-content\") pod \"community-operators-nwgpc\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.938901 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-utilities\") pod \"community-operators-nwgpc\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.938929 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-catalog-content\") pod \"community-operators-nwgpc\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:26 crc kubenswrapper[4719]: I1215 13:52:26.957825 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qccld\" (UniqueName: \"kubernetes.io/projected/c5111a80-08c9-4933-8a65-906b58149ade-kube-api-access-qccld\") pod \"community-operators-nwgpc\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:27 crc kubenswrapper[4719]: I1215 13:52:27.130804 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:27 crc kubenswrapper[4719]: I1215 13:52:27.857721 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwgpc"] Dec 15 13:52:28 crc kubenswrapper[4719]: I1215 13:52:28.616966 4719 generic.go:334] "Generic (PLEG): container finished" podID="c5111a80-08c9-4933-8a65-906b58149ade" containerID="bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5" exitCode=0 Dec 15 13:52:28 crc kubenswrapper[4719]: I1215 13:52:28.617494 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwgpc" event={"ID":"c5111a80-08c9-4933-8a65-906b58149ade","Type":"ContainerDied","Data":"bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5"} Dec 15 13:52:28 crc kubenswrapper[4719]: I1215 13:52:28.617539 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwgpc" event={"ID":"c5111a80-08c9-4933-8a65-906b58149ade","Type":"ContainerStarted","Data":"f30e24864b536da65eb9089f3a4cbf4daa460c57bffbe0fbf60d434d54fcfeea"} Dec 15 13:52:28 crc kubenswrapper[4719]: I1215 13:52:28.621045 4719 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 15 13:52:30 crc kubenswrapper[4719]: I1215 13:52:30.637399 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwgpc" event={"ID":"c5111a80-08c9-4933-8a65-906b58149ade","Type":"ContainerStarted","Data":"d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe"} Dec 15 13:52:31 crc kubenswrapper[4719]: I1215 13:52:31.646260 4719 generic.go:334] "Generic (PLEG): container finished" podID="c5111a80-08c9-4933-8a65-906b58149ade" containerID="d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe" exitCode=0 Dec 15 13:52:31 crc kubenswrapper[4719]: I1215 13:52:31.646417 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwgpc" event={"ID":"c5111a80-08c9-4933-8a65-906b58149ade","Type":"ContainerDied","Data":"d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe"} Dec 15 13:52:32 crc kubenswrapper[4719]: I1215 13:52:32.657308 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwgpc" event={"ID":"c5111a80-08c9-4933-8a65-906b58149ade","Type":"ContainerStarted","Data":"12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0"} Dec 15 13:52:32 crc kubenswrapper[4719]: I1215 13:52:32.681995 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nwgpc" podStartSLOduration=3.158783599 podStartE2EDuration="6.681978126s" podCreationTimestamp="2025-12-15 13:52:26 +0000 UTC" firstStartedPulling="2025-12-15 13:52:28.620676454 +0000 UTC m=+5709.562969484" lastFinishedPulling="2025-12-15 13:52:32.143870981 +0000 UTC m=+5713.086164011" observedRunningTime="2025-12-15 13:52:32.676285517 +0000 UTC m=+5713.618578537" watchObservedRunningTime="2025-12-15 13:52:32.681978126 +0000 UTC m=+5713.624271166" Dec 15 13:52:37 crc kubenswrapper[4719]: I1215 13:52:37.132416 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:37 crc kubenswrapper[4719]: I1215 13:52:37.132830 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:37 crc kubenswrapper[4719]: I1215 13:52:37.175633 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:37 crc kubenswrapper[4719]: I1215 13:52:37.756375 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:37 crc kubenswrapper[4719]: I1215 13:52:37.830847 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwgpc"] Dec 15 13:52:39 crc kubenswrapper[4719]: I1215 13:52:39.713655 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nwgpc" podUID="c5111a80-08c9-4933-8a65-906b58149ade" containerName="registry-server" containerID="cri-o://12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0" gracePeriod=2 Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.181717 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.216102 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qccld\" (UniqueName: \"kubernetes.io/projected/c5111a80-08c9-4933-8a65-906b58149ade-kube-api-access-qccld\") pod \"c5111a80-08c9-4933-8a65-906b58149ade\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.216175 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-catalog-content\") pod \"c5111a80-08c9-4933-8a65-906b58149ade\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.216305 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-utilities\") pod \"c5111a80-08c9-4933-8a65-906b58149ade\" (UID: \"c5111a80-08c9-4933-8a65-906b58149ade\") " Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.217688 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-utilities" (OuterVolumeSpecName: "utilities") pod "c5111a80-08c9-4933-8a65-906b58149ade" (UID: "c5111a80-08c9-4933-8a65-906b58149ade"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.223077 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5111a80-08c9-4933-8a65-906b58149ade-kube-api-access-qccld" (OuterVolumeSpecName: "kube-api-access-qccld") pod "c5111a80-08c9-4933-8a65-906b58149ade" (UID: "c5111a80-08c9-4933-8a65-906b58149ade"). InnerVolumeSpecName "kube-api-access-qccld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.285429 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5111a80-08c9-4933-8a65-906b58149ade" (UID: "c5111a80-08c9-4933-8a65-906b58149ade"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.318350 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qccld\" (UniqueName: \"kubernetes.io/projected/c5111a80-08c9-4933-8a65-906b58149ade-kube-api-access-qccld\") on node \"crc\" DevicePath \"\"" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.318391 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.318400 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5111a80-08c9-4933-8a65-906b58149ade-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.723116 4719 generic.go:334] "Generic (PLEG): container finished" podID="c5111a80-08c9-4933-8a65-906b58149ade" containerID="12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0" exitCode=0 Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.723185 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwgpc" event={"ID":"c5111a80-08c9-4933-8a65-906b58149ade","Type":"ContainerDied","Data":"12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0"} Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.723255 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwgpc" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.723303 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwgpc" event={"ID":"c5111a80-08c9-4933-8a65-906b58149ade","Type":"ContainerDied","Data":"f30e24864b536da65eb9089f3a4cbf4daa460c57bffbe0fbf60d434d54fcfeea"} Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.723333 4719 scope.go:117] "RemoveContainer" containerID="12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.752320 4719 scope.go:117] "RemoveContainer" containerID="d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.775630 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwgpc"] Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.784842 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nwgpc"] Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.786297 4719 scope.go:117] "RemoveContainer" containerID="bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.843679 4719 scope.go:117] "RemoveContainer" containerID="12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0" Dec 15 13:52:40 crc kubenswrapper[4719]: E1215 13:52:40.853161 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0\": container with ID starting with 12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0 not found: ID does not exist" containerID="12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.853211 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0"} err="failed to get container status \"12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0\": rpc error: code = NotFound desc = could not find container \"12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0\": container with ID starting with 12d8097656b7be79ea571f487283b1c4d217705c294e121d9ac714591a9c6bb0 not found: ID does not exist" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.853244 4719 scope.go:117] "RemoveContainer" containerID="d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe" Dec 15 13:52:40 crc kubenswrapper[4719]: E1215 13:52:40.853474 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe\": container with ID starting with d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe not found: ID does not exist" containerID="d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.853501 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe"} err="failed to get container status \"d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe\": rpc error: code = NotFound desc = could not find container \"d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe\": container with ID starting with d9b48684f6ad6bc7988424ae73f22895860e5c5ea66ad1892f5ae2f1baa7cabe not found: ID does not exist" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.853518 4719 scope.go:117] "RemoveContainer" containerID="bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5" Dec 15 13:52:40 crc kubenswrapper[4719]: E1215 13:52:40.853724 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5\": container with ID starting with bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5 not found: ID does not exist" containerID="bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5" Dec 15 13:52:40 crc kubenswrapper[4719]: I1215 13:52:40.853748 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5"} err="failed to get container status \"bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5\": rpc error: code = NotFound desc = could not find container \"bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5\": container with ID starting with bcd4bbfd605174d24f882e3bc6d10728b2659fa4e6d83755f2b508928f2b5ea5 not found: ID does not exist" Dec 15 13:52:41 crc kubenswrapper[4719]: I1215 13:52:41.499373 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5111a80-08c9-4933-8a65-906b58149ade" path="/var/lib/kubelet/pods/c5111a80-08c9-4933-8a65-906b58149ade/volumes" Dec 15 13:52:51 crc kubenswrapper[4719]: I1215 13:52:51.356917 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:52:51 crc kubenswrapper[4719]: I1215 13:52:51.357451 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:53:21 crc kubenswrapper[4719]: I1215 13:53:21.359565 4719 patch_prober.go:28] interesting pod/machine-config-daemon-ff9fm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 15 13:53:21 crc kubenswrapper[4719]: I1215 13:53:21.361450 4719 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 15 13:53:21 crc kubenswrapper[4719]: I1215 13:53:21.361597 4719 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" Dec 15 13:53:21 crc kubenswrapper[4719]: I1215 13:53:21.362479 4719 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7"} pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 15 13:53:21 crc kubenswrapper[4719]: I1215 13:53:21.362675 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerName="machine-config-daemon" containerID="cri-o://7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" gracePeriod=600 Dec 15 13:53:21 crc kubenswrapper[4719]: E1215 13:53:21.495232 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:53:22 crc kubenswrapper[4719]: I1215 13:53:22.076103 4719 generic.go:334] "Generic (PLEG): container finished" podID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" exitCode=0 Dec 15 13:53:22 crc kubenswrapper[4719]: I1215 13:53:22.076210 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" event={"ID":"eaba47bc-655f-4987-b3c6-b8d49ad5c128","Type":"ContainerDied","Data":"7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7"} Dec 15 13:53:22 crc kubenswrapper[4719]: I1215 13:53:22.076483 4719 scope.go:117] "RemoveContainer" containerID="d07a1864400809645dee53e0d4f9f1e7b25073a20cbc7b6da401e405d87bc912" Dec 15 13:53:22 crc kubenswrapper[4719]: I1215 13:53:22.077272 4719 scope.go:117] "RemoveContainer" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" Dec 15 13:53:22 crc kubenswrapper[4719]: E1215 13:53:22.077514 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:53:35 crc kubenswrapper[4719]: I1215 13:53:35.476637 4719 scope.go:117] "RemoveContainer" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" Dec 15 13:53:35 crc kubenswrapper[4719]: E1215 13:53:35.477489 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:53:50 crc kubenswrapper[4719]: I1215 13:53:50.476344 4719 scope.go:117] "RemoveContainer" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" Dec 15 13:53:50 crc kubenswrapper[4719]: E1215 13:53:50.477127 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:54:01 crc kubenswrapper[4719]: I1215 13:54:01.477540 4719 scope.go:117] "RemoveContainer" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" Dec 15 13:54:01 crc kubenswrapper[4719]: E1215 13:54:01.478713 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:54:13 crc kubenswrapper[4719]: I1215 13:54:13.476244 4719 scope.go:117] "RemoveContainer" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" Dec 15 13:54:13 crc kubenswrapper[4719]: E1215 13:54:13.477088 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.195764 4719 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bcjpf"] Dec 15 13:54:18 crc kubenswrapper[4719]: E1215 13:54:18.196543 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5111a80-08c9-4933-8a65-906b58149ade" containerName="extract-content" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.196560 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5111a80-08c9-4933-8a65-906b58149ade" containerName="extract-content" Dec 15 13:54:18 crc kubenswrapper[4719]: E1215 13:54:18.196585 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5111a80-08c9-4933-8a65-906b58149ade" containerName="registry-server" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.196593 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5111a80-08c9-4933-8a65-906b58149ade" containerName="registry-server" Dec 15 13:54:18 crc kubenswrapper[4719]: E1215 13:54:18.196622 4719 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5111a80-08c9-4933-8a65-906b58149ade" containerName="extract-utilities" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.196631 4719 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5111a80-08c9-4933-8a65-906b58149ade" containerName="extract-utilities" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.196884 4719 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5111a80-08c9-4933-8a65-906b58149ade" containerName="registry-server" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.198212 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.266122 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bcjpf"] Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.320379 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-catalog-content\") pod \"redhat-operators-bcjpf\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.320472 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r8dd\" (UniqueName: \"kubernetes.io/projected/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-kube-api-access-5r8dd\") pod \"redhat-operators-bcjpf\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.320781 4719 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-utilities\") pod \"redhat-operators-bcjpf\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.423202 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-catalog-content\") pod \"redhat-operators-bcjpf\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.423278 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r8dd\" (UniqueName: \"kubernetes.io/projected/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-kube-api-access-5r8dd\") pod \"redhat-operators-bcjpf\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.423382 4719 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-utilities\") pod \"redhat-operators-bcjpf\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.423748 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-catalog-content\") pod \"redhat-operators-bcjpf\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.423783 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-utilities\") pod \"redhat-operators-bcjpf\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.447476 4719 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r8dd\" (UniqueName: \"kubernetes.io/projected/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-kube-api-access-5r8dd\") pod \"redhat-operators-bcjpf\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:18 crc kubenswrapper[4719]: I1215 13:54:18.521912 4719 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:19 crc kubenswrapper[4719]: I1215 13:54:19.063629 4719 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bcjpf"] Dec 15 13:54:19 crc kubenswrapper[4719]: I1215 13:54:19.860174 4719 generic.go:334] "Generic (PLEG): container finished" podID="0497f8b1-79eb-43cd-9ac4-6b02e27b5159" containerID="a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9" exitCode=0 Dec 15 13:54:19 crc kubenswrapper[4719]: I1215 13:54:19.860336 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcjpf" event={"ID":"0497f8b1-79eb-43cd-9ac4-6b02e27b5159","Type":"ContainerDied","Data":"a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9"} Dec 15 13:54:19 crc kubenswrapper[4719]: I1215 13:54:19.860497 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcjpf" event={"ID":"0497f8b1-79eb-43cd-9ac4-6b02e27b5159","Type":"ContainerStarted","Data":"d9fb23009418d93e6ab51d0fa15fd7b84d5afd2c352810ed0d97011ccd7781f2"} Dec 15 13:54:21 crc kubenswrapper[4719]: I1215 13:54:21.881553 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcjpf" event={"ID":"0497f8b1-79eb-43cd-9ac4-6b02e27b5159","Type":"ContainerStarted","Data":"f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90"} Dec 15 13:54:24 crc kubenswrapper[4719]: I1215 13:54:24.907672 4719 generic.go:334] "Generic (PLEG): container finished" podID="0497f8b1-79eb-43cd-9ac4-6b02e27b5159" containerID="f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90" exitCode=0 Dec 15 13:54:24 crc kubenswrapper[4719]: I1215 13:54:24.907742 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcjpf" event={"ID":"0497f8b1-79eb-43cd-9ac4-6b02e27b5159","Type":"ContainerDied","Data":"f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90"} Dec 15 13:54:26 crc kubenswrapper[4719]: I1215 13:54:26.934186 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcjpf" event={"ID":"0497f8b1-79eb-43cd-9ac4-6b02e27b5159","Type":"ContainerStarted","Data":"adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248"} Dec 15 13:54:26 crc kubenswrapper[4719]: I1215 13:54:26.965280 4719 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bcjpf" podStartSLOduration=2.944490851 podStartE2EDuration="8.965262295s" podCreationTimestamp="2025-12-15 13:54:18 +0000 UTC" firstStartedPulling="2025-12-15 13:54:19.861813448 +0000 UTC m=+5820.804106478" lastFinishedPulling="2025-12-15 13:54:25.882584892 +0000 UTC m=+5826.824877922" observedRunningTime="2025-12-15 13:54:26.963871311 +0000 UTC m=+5827.906164341" watchObservedRunningTime="2025-12-15 13:54:26.965262295 +0000 UTC m=+5827.907555325" Dec 15 13:54:27 crc kubenswrapper[4719]: I1215 13:54:27.476843 4719 scope.go:117] "RemoveContainer" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" Dec 15 13:54:27 crc kubenswrapper[4719]: E1215 13:54:27.477467 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:54:28 crc kubenswrapper[4719]: I1215 13:54:28.522561 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:28 crc kubenswrapper[4719]: I1215 13:54:28.522606 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:29 crc kubenswrapper[4719]: I1215 13:54:29.574779 4719 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bcjpf" podUID="0497f8b1-79eb-43cd-9ac4-6b02e27b5159" containerName="registry-server" probeResult="failure" output=< Dec 15 13:54:29 crc kubenswrapper[4719]: timeout: failed to connect service ":50051" within 1s Dec 15 13:54:29 crc kubenswrapper[4719]: > Dec 15 13:54:38 crc kubenswrapper[4719]: I1215 13:54:38.584543 4719 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:38 crc kubenswrapper[4719]: I1215 13:54:38.689503 4719 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:38 crc kubenswrapper[4719]: I1215 13:54:38.822070 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bcjpf"] Dec 15 13:54:39 crc kubenswrapper[4719]: I1215 13:54:39.484238 4719 scope.go:117] "RemoveContainer" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" Dec 15 13:54:39 crc kubenswrapper[4719]: E1215 13:54:39.484494 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.037778 4719 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bcjpf" podUID="0497f8b1-79eb-43cd-9ac4-6b02e27b5159" containerName="registry-server" containerID="cri-o://adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248" gracePeriod=2 Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.486182 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.652329 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r8dd\" (UniqueName: \"kubernetes.io/projected/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-kube-api-access-5r8dd\") pod \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.652432 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-catalog-content\") pod \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.652573 4719 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-utilities\") pod \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\" (UID: \"0497f8b1-79eb-43cd-9ac4-6b02e27b5159\") " Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.653714 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-utilities" (OuterVolumeSpecName: "utilities") pod "0497f8b1-79eb-43cd-9ac4-6b02e27b5159" (UID: "0497f8b1-79eb-43cd-9ac4-6b02e27b5159"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.668138 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-kube-api-access-5r8dd" (OuterVolumeSpecName: "kube-api-access-5r8dd") pod "0497f8b1-79eb-43cd-9ac4-6b02e27b5159" (UID: "0497f8b1-79eb-43cd-9ac4-6b02e27b5159"). InnerVolumeSpecName "kube-api-access-5r8dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.755546 4719 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r8dd\" (UniqueName: \"kubernetes.io/projected/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-kube-api-access-5r8dd\") on node \"crc\" DevicePath \"\"" Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.755576 4719 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-utilities\") on node \"crc\" DevicePath \"\"" Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.763891 4719 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0497f8b1-79eb-43cd-9ac4-6b02e27b5159" (UID: "0497f8b1-79eb-43cd-9ac4-6b02e27b5159"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 15 13:54:40 crc kubenswrapper[4719]: I1215 13:54:40.856921 4719 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0497f8b1-79eb-43cd-9ac4-6b02e27b5159-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.047898 4719 generic.go:334] "Generic (PLEG): container finished" podID="0497f8b1-79eb-43cd-9ac4-6b02e27b5159" containerID="adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248" exitCode=0 Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.047942 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcjpf" event={"ID":"0497f8b1-79eb-43cd-9ac4-6b02e27b5159","Type":"ContainerDied","Data":"adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248"} Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.047973 4719 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bcjpf" event={"ID":"0497f8b1-79eb-43cd-9ac4-6b02e27b5159","Type":"ContainerDied","Data":"d9fb23009418d93e6ab51d0fa15fd7b84d5afd2c352810ed0d97011ccd7781f2"} Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.047997 4719 scope.go:117] "RemoveContainer" containerID="adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.048158 4719 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bcjpf" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.099796 4719 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bcjpf"] Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.101995 4719 scope.go:117] "RemoveContainer" containerID="f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.104390 4719 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bcjpf"] Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.145395 4719 scope.go:117] "RemoveContainer" containerID="a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.188109 4719 scope.go:117] "RemoveContainer" containerID="adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248" Dec 15 13:54:41 crc kubenswrapper[4719]: E1215 13:54:41.188630 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248\": container with ID starting with adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248 not found: ID does not exist" containerID="adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.188666 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248"} err="failed to get container status \"adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248\": rpc error: code = NotFound desc = could not find container \"adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248\": container with ID starting with adb05ca5cf661a17e6714730963be47ef834b0251697face5c67a34698ba5248 not found: ID does not exist" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.188689 4719 scope.go:117] "RemoveContainer" containerID="f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90" Dec 15 13:54:41 crc kubenswrapper[4719]: E1215 13:54:41.189229 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90\": container with ID starting with f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90 not found: ID does not exist" containerID="f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.189256 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90"} err="failed to get container status \"f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90\": rpc error: code = NotFound desc = could not find container \"f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90\": container with ID starting with f75171bd09b1b66431f3e0cc60ac39b6fcc1094189751283122eb35e68c63b90 not found: ID does not exist" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.189271 4719 scope.go:117] "RemoveContainer" containerID="a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9" Dec 15 13:54:41 crc kubenswrapper[4719]: E1215 13:54:41.189577 4719 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9\": container with ID starting with a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9 not found: ID does not exist" containerID="a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.189630 4719 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9"} err="failed to get container status \"a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9\": rpc error: code = NotFound desc = could not find container \"a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9\": container with ID starting with a78d71995aa23ca883a40c9785c9791c0cf7183db0dc7b9f2a3ef44421fac6b9 not found: ID does not exist" Dec 15 13:54:41 crc kubenswrapper[4719]: I1215 13:54:41.489712 4719 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0497f8b1-79eb-43cd-9ac4-6b02e27b5159" path="/var/lib/kubelet/pods/0497f8b1-79eb-43cd-9ac4-6b02e27b5159/volumes" Dec 15 13:54:52 crc kubenswrapper[4719]: I1215 13:54:52.476974 4719 scope.go:117] "RemoveContainer" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" Dec 15 13:54:52 crc kubenswrapper[4719]: E1215 13:54:52.477588 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:55:07 crc kubenswrapper[4719]: I1215 13:55:07.476226 4719 scope.go:117] "RemoveContainer" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" Dec 15 13:55:07 crc kubenswrapper[4719]: E1215 13:55:07.477144 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" Dec 15 13:55:18 crc kubenswrapper[4719]: I1215 13:55:18.476791 4719 scope.go:117] "RemoveContainer" containerID="7e1cbeeacd6bbf5594de98074acd148e6f8db9cfccd104237c6c7ea2124c2be7" Dec 15 13:55:18 crc kubenswrapper[4719]: E1215 13:55:18.477507 4719 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ff9fm_openshift-machine-config-operator(eaba47bc-655f-4987-b3c6-b8d49ad5c128)\"" pod="openshift-machine-config-operator/machine-config-daemon-ff9fm" podUID="eaba47bc-655f-4987-b3c6-b8d49ad5c128" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515120011523024433 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015120011524017351 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015117775520016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015117775520015467 5ustar corecore